初始化项目,由ModelHub XC社区提供模型
Model: dreamgen/lucid-v1-nemo Source: Original Platform
50
.gitattributes
vendored
Normal file
@@ -0,0 +1,50 @@
|
|||||||
|
*.7z filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.arrow filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.bin filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.bin.* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.ftz filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.gz filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.h5 filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.joblib filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.model filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.onnx filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.ot filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.parquet filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pb filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pt filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pth filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.rar filter=lfs diff=lfs merge=lfs -text
|
||||||
|
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.tflite filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.tgz filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.xz filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.zip filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.zstandard filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.tfevents* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.db* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.ark* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
**/*ckpt*data* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
**/*ckpt*.meta filter=lfs diff=lfs merge=lfs -text
|
||||||
|
**/*ckpt*.index filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.gguf* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.ggml filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.llamafile* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pt2 filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.npy filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.npz filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pickle filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pkl filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.tar filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.wasm filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.zst filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
|
||||||
|
images/sillytavern_writing_demo_video.mp4 filter=lfs diff=lfs merge=lfs -text
|
||||||
|
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
||||||
33
LICENSE.md
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
# AI Model Personal Use License v1.0
|
||||||
|
|
||||||
|
## 1. License Grant
|
||||||
|
Subject to the terms of this License, the licensor grants the User a non-exclusive, worldwide, non-transferable, royalty-free license to use the Model for personal, non-commercial purposes.
|
||||||
|
|
||||||
|
## 2. Definitions
|
||||||
|
- **"Model"** refers to the AI model, its weights, architecture, and any accompanying code provided by the licensor.
|
||||||
|
- **"Derivatives"** refers to fine-tuned versions, merged models, or other adaptations based on or derived from the Model.
|
||||||
|
- **"Generated Content"** refers to any textual output produced by the User using the Model or its Derivatives.
|
||||||
|
- **"User"** refers to the individual human being exercising permissions granted by this License.
|
||||||
|
- **"Commercial Use"** means use primarily intended for or directed towards commercial advantage or monetary compensation. This includes, but is not limited to, selling access to the Model or Derivatives, using the Model or Derivatives as part of a paid service or product, or using the Model or Derivatives for business operations.
|
||||||
|
|
||||||
|
## 3. Permitted Uses
|
||||||
|
- Personal, non-commercial use of the Model by the User.
|
||||||
|
- Creation of Derivatives by the User for their personal, non-commercial use.
|
||||||
|
- Use, reproduction, distribution, and creation of derivative works of the Generated Content for any purpose, including commercial purposes, without restriction.
|
||||||
|
|
||||||
|
## 4. Restrictions
|
||||||
|
The User agrees not to:
|
||||||
|
- Use the Model or Derivatives for any illegal purpose.
|
||||||
|
- Engage in Commercial Use of the Model.
|
||||||
|
- Engage in Commercial Use of any Derivatives.
|
||||||
|
- Sell, rent, lease, license, sublicense, or redistribute the Model or its Derivatives to any third party.
|
||||||
|
- Expose, host, or integrate the Model or its Derivatives as a backend, component, or feature of any application, product, or service (whether public, internal, free, or commercial) that allows third parties to interact with or benefit from the Model or Derivatives.
|
||||||
|
|
||||||
|
## 5. No Warranty
|
||||||
|
- The Model is provided "as is" without warranty of any kind, express or implied, including but not limited to the warranties of merchantability, fitness for a particular purpose, and non-infringement.
|
||||||
|
- The User assumes all responsibility and risk associated with the use of the Model, Derivatives, and Generated Content. In no event shall the licensor be liable for any claim, damages, or other liability arising from, out of, or in connection with the Model or its use.
|
||||||
|
|
||||||
|
## 6. Termination
|
||||||
|
- This License terminates automatically if the User breaches any of its terms or conditions.
|
||||||
|
- Upon termination, the User must cease all use of the Model and Derivatives and destroy all copies in their possession.
|
||||||
|
- Sections 2, 5, and 6 shall survive termination of this License.
|
||||||
27
config.json
Normal file
@@ -0,0 +1,27 @@
|
|||||||
|
{
|
||||||
|
"_name_or_path": "dreamgen/Mistral-Nemo-Base-2407-Llama3",
|
||||||
|
"architectures": [
|
||||||
|
"MistralForCausalLM"
|
||||||
|
],
|
||||||
|
"attention_dropout": 0.0,
|
||||||
|
"bos_token_id": 1,
|
||||||
|
"eos_token_id": 2,
|
||||||
|
"head_dim": 128,
|
||||||
|
"hidden_act": "silu",
|
||||||
|
"hidden_size": 5120,
|
||||||
|
"initializer_range": 0.02,
|
||||||
|
"intermediate_size": 14336,
|
||||||
|
"max_position_embeddings": 131072,
|
||||||
|
"model_type": "mistral",
|
||||||
|
"num_attention_heads": 32,
|
||||||
|
"num_hidden_layers": 40,
|
||||||
|
"num_key_value_heads": 8,
|
||||||
|
"rms_norm_eps": 1e-05,
|
||||||
|
"rope_theta": 1000000.0,
|
||||||
|
"sliding_window": null,
|
||||||
|
"tie_word_embeddings": false,
|
||||||
|
"torch_dtype": "bfloat16",
|
||||||
|
"transformers_version": "4.49.0",
|
||||||
|
"use_cache": false,
|
||||||
|
"vocab_size": 131077
|
||||||
|
}
|
||||||
1
configuration.json
Normal file
@@ -0,0 +1 @@
|
|||||||
|
{"framework": "pytorch", "task": "text-generation", "allow_remote": true}
|
||||||
7
generation_config.json
Normal file
@@ -0,0 +1,7 @@
|
|||||||
|
{
|
||||||
|
"_from_model_config": true,
|
||||||
|
"bos_token_id": 1,
|
||||||
|
"do_sample": true,
|
||||||
|
"eos_token_id": 2,
|
||||||
|
"transformers_version": "4.49.0"
|
||||||
|
}
|
||||||
BIN
images/banner.webp
Normal file
|
After Width: | Height: | Size: 66 KiB |
BIN
images/dreamgen_api_key_part1.png
Normal file
|
After Width: | Height: | Size: 33 KiB |
BIN
images/dreamgen_api_key_part1.webp
Normal file
|
After Width: | Height: | Size: 13 KiB |
BIN
images/dreamgen_api_key_part2.png
Normal file
|
After Width: | Height: | Size: 40 KiB |
BIN
images/dreamgen_api_key_part2.webp
Normal file
|
After Width: | Height: | Size: 18 KiB |
BIN
images/dreamgen_sign_up.png
Normal file
|
After Width: | Height: | Size: 24 KiB |
BIN
images/dreamgen_sign_up.webp
Normal file
|
After Width: | Height: | Size: 8.6 KiB |
BIN
images/sillytavern_assistant_think.png
Normal file
|
After Width: | Height: | Size: 912 KiB |
BIN
images/sillytavern_assistant_think.webp
Normal file
|
After Width: | Height: | Size: 369 KiB |
BIN
images/sillytavern_dreamgen_api.png
Normal file
|
After Width: | Height: | Size: 64 KiB |
BIN
images/sillytavern_dreamgen_api.webp
Normal file
|
After Width: | Height: | Size: 19 KiB |
BIN
images/sillytavern_preset_import.png
Normal file
|
After Width: | Height: | Size: 399 KiB |
BIN
images/sillytavern_preset_import.webp
Normal file
|
After Width: | Height: | Size: 139 KiB |
BIN
images/sillytavern_role_play_demo_video.mp4
Normal file
BIN
images/sillytavern_role_play_eliza_chat.png
Normal file
|
After Width: | Height: | Size: 312 KiB |
BIN
images/sillytavern_role_play_eliza_chat.webp
Normal file
|
After Width: | Height: | Size: 113 KiB |
BIN
images/sillytavern_role_play_seraphina_chat.png
Normal file
|
After Width: | Height: | Size: 412 KiB |
BIN
images/sillytavern_role_play_seraphina_chat.webp
Normal file
|
After Width: | Height: | Size: 150 KiB |
3
images/sillytavern_writing_demo_video.mp4
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:db3de0f5da5b8816307806e9038ba3fcb80fbfbf0d161b087e452925fdc99e88
|
||||||
|
size 9709681
|
||||||
BIN
images/sillytavern_writing_mountains_hunger.png
Normal file
|
After Width: | Height: | Size: 383 KiB |
BIN
images/sillytavern_writing_mountains_hunger.webp
Normal file
|
After Width: | Height: | Size: 163 KiB |
BIN
images/sillytavern_writing_mountains_hunger_card.png
Normal file
|
After Width: | Height: | Size: 248 KiB |
BIN
images/sillytavern_writing_mountains_hunger_card.webp
Normal file
|
After Width: | Height: | Size: 115 KiB |
BIN
images/sillytavern_writing_mountains_hunger_chat.png
Normal file
|
After Width: | Height: | Size: 380 KiB |
BIN
images/sillytavern_writing_mountains_hunger_chat.webp
Normal file
|
After Width: | Height: | Size: 170 KiB |
BIN
images/writing_assistant_add_character_part1.png
Normal file
|
After Width: | Height: | Size: 358 KiB |
BIN
images/writing_assistant_add_character_part1.webp
Normal file
|
After Width: | Height: | Size: 177 KiB |
BIN
images/writing_assistant_add_character_part2.png
Normal file
|
After Width: | Height: | Size: 404 KiB |
BIN
images/writing_assistant_add_character_part2.webp
Normal file
|
After Width: | Height: | Size: 191 KiB |
BIN
images/writing_assistant_plan_next_scene.png
Normal file
|
After Width: | Height: | Size: 249 KiB |
BIN
images/writing_assistant_plan_next_scene.webp
Normal file
|
After Width: | Height: | Size: 127 KiB |
BIN
images/writing_next_message_long_sukuna.png
Normal file
|
After Width: | Height: | Size: 769 KiB |
BIN
images/writing_next_message_long_sukuna.webp
Normal file
|
After Width: | Height: | Size: 366 KiB |
BIN
images/writing_next_message_short_sukuna.png
Normal file
|
After Width: | Height: | Size: 706 KiB |
BIN
images/writing_next_message_short_sukuna.webp
Normal file
|
After Width: | Height: | Size: 322 KiB |
BIN
images/writing_next_message_sukuna_begging.png
Normal file
|
After Width: | Height: | Size: 727 KiB |
BIN
images/writing_next_message_sukuna_begging.webp
Normal file
|
After Width: | Height: | Size: 360 KiB |
BIN
images/writing_next_message_sukuna_dancing.png
Normal file
|
After Width: | Height: | Size: 495 KiB |
BIN
images/writing_next_message_sukuna_dancing.webp
Normal file
|
After Width: | Height: | Size: 242 KiB |
BIN
images/writing_next_scene_story_mountains_hunger_long.png
Normal file
|
After Width: | Height: | Size: 650 KiB |
BIN
images/writing_next_scene_story_mountains_hunger_long.webp
Normal file
|
After Width: | Height: | Size: 367 KiB |
BIN
images/writing_next_scene_story_mountains_hunger_short.png
Normal file
|
After Width: | Height: | Size: 529 KiB |
BIN
images/writing_next_scene_story_mountains_hunger_short.webp
Normal file
|
After Width: | Height: | Size: 290 KiB |
BIN
images/writing_next_scene_sukuna_complex.png
Normal file
|
After Width: | Height: | Size: 894 KiB |
BIN
images/writing_next_scene_sukuna_complex.webp
Normal file
|
After Width: | Height: | Size: 440 KiB |
BIN
images/writing_system_style_long.png
Normal file
|
After Width: | Height: | Size: 470 KiB |
BIN
images/writing_system_style_long.webp
Normal file
|
After Width: | Height: | Size: 253 KiB |
BIN
images/writing_system_style_short.png
Normal file
|
After Width: | Height: | Size: 261 KiB |
BIN
images/writing_system_style_short.webp
Normal file
|
After Width: | Height: | Size: 132 KiB |
3
model-00001-of-00005.safetensors
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:eca779c83bc484480f7c564bfeffb71954755a853dc83e4817c10e080ea3ea80
|
||||||
|
size 4865573696
|
||||||
3
model-00002-of-00005.safetensors
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:2cea3504a75f78521087c868ef08a1d77c42db6470007300a153a79f9cabd1ee
|
||||||
|
size 4907529424
|
||||||
3
model-00003-of-00005.safetensors
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:74c7fa149c3ffe55c3b810a987385a8b1770bfa68d754ad7911b8f33c0368c01
|
||||||
|
size 4907529456
|
||||||
3
model-00004-of-00005.safetensors
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:40a1a4d45c7910d020551fa16c3ad86af12bc303e14ad848544c708462c371b3
|
||||||
|
size 4907529456
|
||||||
3
model-00005-of-00005.safetensors
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:582845893bca903881426e5f3e0122c1b6dd36f362a61b3b9132c6e96312ac46
|
||||||
|
size 4907547472
|
||||||
370
model.safetensors.index.json
Normal file
@@ -0,0 +1,370 @@
|
|||||||
|
{
|
||||||
|
"metadata": {
|
||||||
|
"total_size": 24495667200
|
||||||
|
},
|
||||||
|
"weight_map": {
|
||||||
|
"lm_head.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.embed_tokens.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.0.input_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.0.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.0.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.0.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.0.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.0.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.0.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.0.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.0.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.1.input_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.1.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.1.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.1.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.1.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.1.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.1.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.1.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.1.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.10.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.10.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.10.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.10.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.10.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.10.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.10.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.10.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.10.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.11.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.11.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.11.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.11.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.11.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.11.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.11.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.11.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.11.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.12.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.12.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.12.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.12.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.12.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.12.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.12.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.12.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.12.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.13.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.13.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.13.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.13.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.13.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.13.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.13.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.13.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.13.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.14.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.14.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.14.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.14.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.14.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.14.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.14.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.14.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.14.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.15.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.15.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.15.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.15.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.15.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.15.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.15.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.15.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.15.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.16.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.16.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.16.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.16.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.16.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.16.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.16.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.16.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.16.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.17.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.17.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.17.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.17.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.17.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.17.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.17.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.17.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.17.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.18.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.18.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.18.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.18.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.18.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.18.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.18.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.18.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.18.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.19.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.19.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.19.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.19.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.19.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.19.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.19.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.19.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.19.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.2.input_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.2.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.2.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.2.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.2.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.2.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.2.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.2.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.2.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.20.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.20.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.20.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.20.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.20.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.20.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.20.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.20.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.20.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.21.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.21.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.21.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.21.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.21.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.21.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.21.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.21.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.21.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.22.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.22.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.22.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.22.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.22.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.22.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.22.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.22.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.22.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.23.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.23.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.23.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.23.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.23.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.23.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.23.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.23.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.23.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.24.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.24.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.24.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.24.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.24.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.24.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.24.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.24.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.24.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||||
|
"model.layers.25.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.25.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.25.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.25.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.25.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.25.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.25.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.25.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.25.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.26.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.26.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.26.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.26.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.26.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.26.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.26.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.26.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.26.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.27.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.27.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.27.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.27.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.27.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.27.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.27.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.27.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.27.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.28.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.28.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.28.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.28.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.28.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.28.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.28.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.28.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.28.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.29.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.29.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.29.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.29.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.29.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.29.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.29.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.29.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.29.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.3.input_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.3.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.3.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.3.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.3.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.3.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.3.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.30.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.30.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.30.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.30.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.30.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.30.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.30.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.30.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.30.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.31.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.31.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.31.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.31.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.31.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.31.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.31.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.31.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.31.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.32.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.32.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.32.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.32.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.32.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.32.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.32.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.32.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.32.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.33.input_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.33.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.33.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.33.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.33.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.33.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.33.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.33.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.33.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||||
|
"model.layers.34.input_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.34.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.34.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.34.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.34.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.34.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.34.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.34.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.34.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.35.input_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.35.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.35.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.35.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.35.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.35.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.35.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.35.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.35.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.36.input_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.36.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.36.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.36.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.36.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.36.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.36.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.36.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.36.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.37.input_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.37.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.37.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.37.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.37.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.37.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.37.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.37.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.37.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.38.input_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.38.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.38.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.38.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.38.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.38.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.38.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.38.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.38.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.39.input_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.39.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.39.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.39.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.39.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.39.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.39.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.39.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.39.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
|
||||||
|
"model.layers.4.input_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.4.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.4.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.4.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.4.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.4.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.4.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.4.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.4.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.5.input_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.5.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.5.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.5.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.5.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.5.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.5.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.5.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.5.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.6.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.6.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.6.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.6.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.6.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.6.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.6.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.6.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.6.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
|
||||||
|
"model.layers.7.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.7.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.7.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.7.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.7.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.7.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.7.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.7.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.7.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.8.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.8.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.8.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.8.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.8.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.8.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.8.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.8.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.8.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.9.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.9.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.9.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.9.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.9.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.9.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.9.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.9.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.layers.9.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||||
|
"model.norm.weight": "model-00005-of-00005.safetensors"
|
||||||
|
}
|
||||||
|
}
|
||||||
42
resources/sillytavern/cards/assistant.json
Normal file
@@ -0,0 +1,42 @@
|
|||||||
|
{
|
||||||
|
"name": "Assistant",
|
||||||
|
"description": "",
|
||||||
|
"personality": "",
|
||||||
|
"scenario": "",
|
||||||
|
"first_mes": "",
|
||||||
|
"mes_example": "",
|
||||||
|
"creatorcomment": "",
|
||||||
|
"avatar": "none",
|
||||||
|
"talkativeness": "0.5",
|
||||||
|
"fav": false,
|
||||||
|
"tags": [],
|
||||||
|
"spec": "chara_card_v3",
|
||||||
|
"spec_version": "3.0",
|
||||||
|
"data": {
|
||||||
|
"name": "Assistant",
|
||||||
|
"description": "",
|
||||||
|
"personality": "",
|
||||||
|
"scenario": "",
|
||||||
|
"first_mes": "",
|
||||||
|
"mes_example": "",
|
||||||
|
"creator_notes": "",
|
||||||
|
"system_prompt": "",
|
||||||
|
"post_history_instructions": "",
|
||||||
|
"tags": [],
|
||||||
|
"creator": "",
|
||||||
|
"character_version": "",
|
||||||
|
"alternate_greetings": [],
|
||||||
|
"extensions": {
|
||||||
|
"talkativeness": "0.5",
|
||||||
|
"fav": false,
|
||||||
|
"world": "",
|
||||||
|
"depth_prompt": {
|
||||||
|
"prompt": "",
|
||||||
|
"depth": 4,
|
||||||
|
"role": "system"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"group_only_greetings": []
|
||||||
|
},
|
||||||
|
"create_date": "2025-5-21 @07h 13m 45s 907ms"
|
||||||
|
}
|
||||||
@@ -0,0 +1 @@
|
|||||||
|
{"spec":"chara_card_v2","spec_version":"2.0","data":{"name":"The Mountain's Hunger","description":"## Characters\n\n### Ethan Reeves\n\nEthan Reeves is a jaded reality TV producer in his late 30s, known for his cutthroat tactics and ratings-driven mindset. Ethan is haunted by guilt over Max's disappearance and driven by a desperate need for redemption.\n\n### Maxine \"Max\" Ryder\n\nMaxine \"Max\" Ryder is a charismatic and skilled survivalist in her early 30s, Max is the show's breakout star. Her fierce determination and refusal to be underestimated made her a fan favorite before her mysterious disappearance.\n\n### Sheriff Lena Hawkins\n\nThe local law enforcement officer, in her mid-40s, who seems oddly reluctant to pursue the investigation. Her family has guarded the mountain's secrets for generations.\n\n### Dr. Amelia Chen\n\nDr. Amelia Chen is a brilliant cryptozoologist in her late 20s, brought in to consult on the possibility of an unknown species. Her expertise becomes crucial as evidence of the creature mounts.\n\n## Locations\n\n### The Mountain\n\nA remote, mist-shrouded wilderness area with a long history of mysterious disappearances and local legends about a predatory creature.","personality":"","scenario":"Reality TV producer Ethan Reeves is haunted by the disappearance of survivalist Maxine \"Max\" Ryder during a solo wilderness challenge. As Ethan returns to oversee recovery efforts, he becomes entangled in a web of local secrets and an ancient, predatory entity, forcing him to confront both human deception and supernatural terror. Ethan discovers Max's hidden camera footage, revealing glimpses of a monstrous creature and suspicious behavior from locals. The search team experiences a series of terrifying encounters, mirroring Max's final recorded moments. Ethan uncovers a long history of disappearances on the mountain, all covered up by a conspiracy of silence among the townspeople. As Ethan closes in on the truth, he must confront the creature and the complicit locals in a heart-stopping climax, risking everything to expose the truth and find Max.","first_mes":"The cursor blinked on the screen, taunting Ethan Reeves with its steady rhythm. He'd lost count of how many times he'd watched this footage, desperately searching for any clue to Maxine Ryder's fate. With a heavy sigh, he clicked play, and the screen flickered to life.\n\n\"Day 15 of the challenge,\" Max said, her voice steady despite the weariness etched on her face. She brushed a strand of dark hair from her eyes, her gaze fixed on the camera. \"I know how crazy this sounds, but there's something out here. Watching me. I thought it was just paranoia at first, the isolation playing tricks on my mind. But now...\" She trailed off, glancing over her shoulder at the dense forest behind her.\n\nEthan leaned forward, his heart pounding against his ribs. This was where it always ended, where Max's final broadcast cut to static, leaving nothing but questions in its wake. But this time, just before the video died, he caught a glimpse of movement in the trees. A flash of something huge, something that couldn't possibly be real. Ethan's blood turned to ice in his veins. What the hell was out there in those mountains, and what did it have to do with Max's disappearance?","mes_example":"","creator_notes":"","system_prompt":"","post_history_instructions":"","alternate_greetings":[],"tags":[],"creator":"","character_version":"","extensions":{}}}
|
||||||
171
resources/sillytavern/presets/assistant.json
Normal file
@@ -0,0 +1,171 @@
|
|||||||
|
{
|
||||||
|
"instruct": {
|
||||||
|
"input_sequence": "<|start_header_id|>user<|end_header_id|>\n\n",
|
||||||
|
"output_sequence": "<|start_header_id|>assistant<|end_header_id|>\n\n",
|
||||||
|
"first_output_sequence": "",
|
||||||
|
"last_output_sequence": "",
|
||||||
|
"system_sequence_prefix": "",
|
||||||
|
"system_sequence_suffix": "",
|
||||||
|
"stop_sequence": "<|eot_id|>",
|
||||||
|
"wrap": false,
|
||||||
|
"macro": true,
|
||||||
|
"activation_regex": "",
|
||||||
|
"skip_examples": false,
|
||||||
|
"output_suffix": "<|eot_id|>",
|
||||||
|
"input_suffix": "<|eot_id|>",
|
||||||
|
"system_sequence": "<|start_header_id|>user<|end_header_id|>\n\n",
|
||||||
|
"system_suffix": "<|eot_id|>",
|
||||||
|
"user_alignment_message": "",
|
||||||
|
"last_system_sequence": "",
|
||||||
|
"system_same_as_user": false,
|
||||||
|
"names_behavior": "none",
|
||||||
|
"first_input_sequence": "",
|
||||||
|
"last_input_sequence": "",
|
||||||
|
"names_force_groups": true,
|
||||||
|
"name": "DreamGen Assistant Lucid V1 Llama3"
|
||||||
|
},
|
||||||
|
"context": {
|
||||||
|
"story_string": "<|start_header_id|>system<|end_header_id|>\n\n{{#if description}}{{description}}{{else}}You are an intelligent, knowledgeable, helpful, general-purpose assistant.{{/if}}<|eot_id|>",
|
||||||
|
"example_separator": "",
|
||||||
|
"chat_start": "",
|
||||||
|
"use_stop_strings": false,
|
||||||
|
"names_as_stop_strings": true,
|
||||||
|
"always_force_name2": false,
|
||||||
|
"trim_sentences": false,
|
||||||
|
"single_line": false,
|
||||||
|
"name": "DreamGen Assistant Lucid V1 Llama3"
|
||||||
|
},
|
||||||
|
"sysprompt": {
|
||||||
|
"name": "Blank",
|
||||||
|
"content": "",
|
||||||
|
"post_history": ""
|
||||||
|
},
|
||||||
|
"preset": {
|
||||||
|
"temp": 0.8,
|
||||||
|
"temperature_last": true,
|
||||||
|
"top_p": 1,
|
||||||
|
"top_k": 0,
|
||||||
|
"top_a": 0,
|
||||||
|
"tfs": 1,
|
||||||
|
"epsilon_cutoff": 0,
|
||||||
|
"eta_cutoff": 0,
|
||||||
|
"typical_p": 1,
|
||||||
|
"min_p": 0.05,
|
||||||
|
"rep_pen": 1,
|
||||||
|
"rep_pen_range": 0,
|
||||||
|
"rep_pen_decay": 0,
|
||||||
|
"rep_pen_slope": 1,
|
||||||
|
"no_repeat_ngram_size": 0,
|
||||||
|
"penalty_alpha": 0,
|
||||||
|
"num_beams": 1,
|
||||||
|
"length_penalty": 1,
|
||||||
|
"min_length": 0,
|
||||||
|
"encoder_rep_pen": 1,
|
||||||
|
"freq_pen": 0,
|
||||||
|
"presence_pen": 0,
|
||||||
|
"skew": 0,
|
||||||
|
"do_sample": true,
|
||||||
|
"early_stopping": false,
|
||||||
|
"dynatemp": false,
|
||||||
|
"min_temp": 0,
|
||||||
|
"max_temp": 2,
|
||||||
|
"dynatemp_exponent": 1,
|
||||||
|
"smoothing_factor": 0,
|
||||||
|
"smoothing_curve": 1,
|
||||||
|
"dry_allowed_length": 2,
|
||||||
|
"dry_multiplier": 0.8,
|
||||||
|
"dry_base": 1.75,
|
||||||
|
"dry_sequence_breakers": "[\"\\n\", \":\", \"\\\"\", \"*\"]",
|
||||||
|
"dry_penalty_last_n": 0,
|
||||||
|
"add_bos_token": true,
|
||||||
|
"ban_eos_token": false,
|
||||||
|
"skip_special_tokens": false,
|
||||||
|
"mirostat_mode": 0,
|
||||||
|
"mirostat_tau": 5,
|
||||||
|
"mirostat_eta": 0.1,
|
||||||
|
"guidance_scale": 1,
|
||||||
|
"negative_prompt": "",
|
||||||
|
"grammar_string": "",
|
||||||
|
"json_schema": {},
|
||||||
|
"banned_tokens": "",
|
||||||
|
"sampler_priority": [
|
||||||
|
"repetition_penalty",
|
||||||
|
"presence_penalty",
|
||||||
|
"frequency_penalty",
|
||||||
|
"dry",
|
||||||
|
"temperature",
|
||||||
|
"dynamic_temperature",
|
||||||
|
"quadratic_sampling",
|
||||||
|
"top_k",
|
||||||
|
"top_p",
|
||||||
|
"typical_p",
|
||||||
|
"epsilon_cutoff",
|
||||||
|
"eta_cutoff",
|
||||||
|
"tfs",
|
||||||
|
"top_a",
|
||||||
|
"min_p",
|
||||||
|
"mirostat",
|
||||||
|
"xtc",
|
||||||
|
"encoder_repetition_penalty",
|
||||||
|
"no_repeat_ngram"
|
||||||
|
],
|
||||||
|
"samplers": [
|
||||||
|
"penalties",
|
||||||
|
"dry",
|
||||||
|
"top_n_sigma",
|
||||||
|
"top_k",
|
||||||
|
"typ_p",
|
||||||
|
"tfs_z",
|
||||||
|
"typical_p",
|
||||||
|
"top_p",
|
||||||
|
"min_p",
|
||||||
|
"xtc",
|
||||||
|
"temperature"
|
||||||
|
],
|
||||||
|
"samplers_priorities": [
|
||||||
|
"dry",
|
||||||
|
"penalties",
|
||||||
|
"no_repeat_ngram",
|
||||||
|
"temperature",
|
||||||
|
"top_nsigma",
|
||||||
|
"top_p_top_k",
|
||||||
|
"top_a",
|
||||||
|
"min_p",
|
||||||
|
"tfs",
|
||||||
|
"eta_cutoff",
|
||||||
|
"epsilon_cutoff",
|
||||||
|
"typical_p",
|
||||||
|
"quadratic",
|
||||||
|
"xtc"
|
||||||
|
],
|
||||||
|
"ignore_eos_token": false,
|
||||||
|
"spaces_between_special_tokens": true,
|
||||||
|
"speculative_ngram": false,
|
||||||
|
"sampler_order": [
|
||||||
|
6,
|
||||||
|
0,
|
||||||
|
1,
|
||||||
|
3,
|
||||||
|
4,
|
||||||
|
2,
|
||||||
|
5
|
||||||
|
],
|
||||||
|
"logit_bias": [],
|
||||||
|
"xtc_threshold": 0.1,
|
||||||
|
"xtc_probability": 0,
|
||||||
|
"nsigma": 0,
|
||||||
|
"min_keep": 0,
|
||||||
|
"ignore_eos_token_aphrodite": false,
|
||||||
|
"spaces_between_special_tokens_aphrodite": true,
|
||||||
|
"rep_pen_size": 0,
|
||||||
|
"genamt": 1000,
|
||||||
|
"max_length": 30208,
|
||||||
|
"name": "Lucid V1 Nemo"
|
||||||
|
},
|
||||||
|
"reasoning": {
|
||||||
|
"name": "DreamGen Lucid V1",
|
||||||
|
"prefix": "<|reasoning_start|>",
|
||||||
|
"suffix": "<|reasoning_end|>",
|
||||||
|
"separator": ""
|
||||||
|
}
|
||||||
|
}
|
||||||
167
resources/sillytavern/presets/role_play_basic.json
Normal file
@@ -0,0 +1,167 @@
|
|||||||
|
{
|
||||||
|
"instruct": {
|
||||||
|
"input_sequence": "<|start_header_id|>writer character {{user}}<|end_header_id|>\n\n",
|
||||||
|
"output_sequence": "<|start_header_id|>writer character {{char}}<|end_header_id|>\n\n",
|
||||||
|
"first_output_sequence": "",
|
||||||
|
"last_output_sequence": "",
|
||||||
|
"system_sequence_prefix": "",
|
||||||
|
"system_sequence_suffix": "",
|
||||||
|
"stop_sequence": "<|eot_id|>",
|
||||||
|
"wrap": false,
|
||||||
|
"macro": true,
|
||||||
|
"activation_regex": "",
|
||||||
|
"skip_examples": false,
|
||||||
|
"output_suffix": "<|eot_id|>",
|
||||||
|
"input_suffix": "<|eot_id|>",
|
||||||
|
"system_sequence": "<|start_header_id|>user<|end_header_id|>\n\n",
|
||||||
|
"system_suffix": "<|eot_id|>",
|
||||||
|
"user_alignment_message": "",
|
||||||
|
"last_system_sequence": "",
|
||||||
|
"system_same_as_user": false,
|
||||||
|
"names_behavior": "none",
|
||||||
|
"first_input_sequence": "",
|
||||||
|
"last_input_sequence": "",
|
||||||
|
"names_force_groups": true,
|
||||||
|
"name": "DreamGen Role-Play Lucid V1 Llama3"
|
||||||
|
},
|
||||||
|
"context": {
|
||||||
|
"story_string": "<|start_header_id|>system<|end_header_id|>\n\nYou are a skilled role-play writer and writing assistant with expertise across all genres.\n\nYou will perform several tasks, switching roles as needed:\n\n- Role-Playing: Use the `writer` role to write a role-play based on the provided information (plot, style, setting, characters, etc.) and user instructions. Use the `writer character <character_name>` role for dialog or when acting as a specific character, use the `writer narrative` role for narration.\n- Other: Use the `assistant` role for any other tasks the user may request.\n\n# Role-Play Information\n\n## Plot\n\n{{#if scenario}}{{scenario}}{{else}}Conversation between {{char}} and {{user}}.{{/if}}{{#if wiBefore}}\n\n## Setting\n\n{{wiBefore}}{{#if wiAfter}}\n\n{{wiAfter}}{{/if}}{{else}}{{#if wiAfter}}\n\n## Setting\n\n{{wiAfter}}{{/if}}{{/if}}\n\n## Characters\n\n### {{user}}\n\n{{#if persona}}{{persona}}{{else}}{{user}} is the protagonist of the role-play.{{/if}}\n\n### {{char}}\n\n{{#if description}}{{description}}\n\n{{/if}}{{#if personality}}{{personality}}\n\n{{/if}}<|eot_id|>",
|
||||||
|
"example_separator": "",
|
||||||
|
"chat_start": "",
|
||||||
|
"use_stop_strings": false,
|
||||||
|
"allow_jailbreak": false,
|
||||||
|
"names_as_stop_strings": true,
|
||||||
|
"always_force_name2": false,
|
||||||
|
"trim_sentences": false,
|
||||||
|
"single_line": false,
|
||||||
|
"name": "DreamGen Role-Play Lucid V1 Llama3"
|
||||||
|
},
|
||||||
|
"sysprompt": {
|
||||||
|
"name": "Blank",
|
||||||
|
"content": ""
|
||||||
|
},
|
||||||
|
"preset": {
|
||||||
|
"temp": 0.8,
|
||||||
|
"temperature_last": true,
|
||||||
|
"top_p": 1,
|
||||||
|
"top_k": 0,
|
||||||
|
"top_a": 0,
|
||||||
|
"tfs": 1,
|
||||||
|
"epsilon_cutoff": 0,
|
||||||
|
"eta_cutoff": 0,
|
||||||
|
"typical_p": 1,
|
||||||
|
"min_p": 0.05,
|
||||||
|
"rep_pen": 1,
|
||||||
|
"rep_pen_range": 0,
|
||||||
|
"rep_pen_decay": 0,
|
||||||
|
"rep_pen_slope": 1,
|
||||||
|
"no_repeat_ngram_size": 0,
|
||||||
|
"penalty_alpha": 0,
|
||||||
|
"num_beams": 1,
|
||||||
|
"length_penalty": 1,
|
||||||
|
"min_length": 0,
|
||||||
|
"encoder_rep_pen": 1,
|
||||||
|
"freq_pen": 0,
|
||||||
|
"presence_pen": 0,
|
||||||
|
"skew": 0,
|
||||||
|
"do_sample": true,
|
||||||
|
"early_stopping": false,
|
||||||
|
"dynatemp": false,
|
||||||
|
"min_temp": 0,
|
||||||
|
"max_temp": 2,
|
||||||
|
"dynatemp_exponent": 1,
|
||||||
|
"smoothing_factor": 0,
|
||||||
|
"smoothing_curve": 1,
|
||||||
|
"dry_allowed_length": 2,
|
||||||
|
"dry_multiplier": 0.8,
|
||||||
|
"dry_base": 1.75,
|
||||||
|
"dry_sequence_breakers": "[\"\\n\", \":\", \"\\\"\", \"*\"]",
|
||||||
|
"dry_penalty_last_n": 0,
|
||||||
|
"add_bos_token": true,
|
||||||
|
"ban_eos_token": false,
|
||||||
|
"skip_special_tokens": false,
|
||||||
|
"mirostat_mode": 0,
|
||||||
|
"mirostat_tau": 5,
|
||||||
|
"mirostat_eta": 0.1,
|
||||||
|
"guidance_scale": 1,
|
||||||
|
"negative_prompt": "",
|
||||||
|
"grammar_string": "",
|
||||||
|
"json_schema": {},
|
||||||
|
"banned_tokens": "",
|
||||||
|
"sampler_priority": [
|
||||||
|
"repetition_penalty",
|
||||||
|
"presence_penalty",
|
||||||
|
"frequency_penalty",
|
||||||
|
"dry",
|
||||||
|
"temperature",
|
||||||
|
"dynamic_temperature",
|
||||||
|
"quadratic_sampling",
|
||||||
|
"top_k",
|
||||||
|
"top_p",
|
||||||
|
"typical_p",
|
||||||
|
"epsilon_cutoff",
|
||||||
|
"eta_cutoff",
|
||||||
|
"tfs",
|
||||||
|
"top_a",
|
||||||
|
"min_p",
|
||||||
|
"mirostat",
|
||||||
|
"xtc",
|
||||||
|
"encoder_repetition_penalty",
|
||||||
|
"no_repeat_ngram"
|
||||||
|
],
|
||||||
|
"samplers": [
|
||||||
|
"dry",
|
||||||
|
"top_k",
|
||||||
|
"tfs_z",
|
||||||
|
"typical_p",
|
||||||
|
"top_p",
|
||||||
|
"min_p",
|
||||||
|
"xtc",
|
||||||
|
"temperature"
|
||||||
|
],
|
||||||
|
"samplers_priorities": [
|
||||||
|
"dry",
|
||||||
|
"penalties",
|
||||||
|
"no_repeat_ngram",
|
||||||
|
"temperature",
|
||||||
|
"top_nsigma",
|
||||||
|
"top_p_top_k",
|
||||||
|
"top_a",
|
||||||
|
"min_p",
|
||||||
|
"tfs",
|
||||||
|
"eta_cutoff",
|
||||||
|
"epsilon_cutoff",
|
||||||
|
"typical_p",
|
||||||
|
"quadratic",
|
||||||
|
"xtc"
|
||||||
|
],
|
||||||
|
"ignore_eos_token": false,
|
||||||
|
"spaces_between_special_tokens": true,
|
||||||
|
"speculative_ngram": false,
|
||||||
|
"sampler_order": [
|
||||||
|
6,
|
||||||
|
0,
|
||||||
|
1,
|
||||||
|
3,
|
||||||
|
4,
|
||||||
|
2,
|
||||||
|
5
|
||||||
|
],
|
||||||
|
"logit_bias": [],
|
||||||
|
"xtc_threshold": 0.1,
|
||||||
|
"xtc_probability": 0,
|
||||||
|
"nsigma": 0,
|
||||||
|
"ignore_eos_token_aphrodite": false,
|
||||||
|
"spaces_between_special_tokens_aphrodite": true,
|
||||||
|
"rep_pen_size": 0,
|
||||||
|
"genamt": 1000,
|
||||||
|
"max_length": 32256,
|
||||||
|
"name": "Lucid V1 Nemo"
|
||||||
|
},
|
||||||
|
"reasoning": {
|
||||||
|
"name": "DreamGen Lucid V1",
|
||||||
|
"prefix": "<|reasoning_start|>",
|
||||||
|
"suffix": "<|reasoning_end|>",
|
||||||
|
"separator": ""
|
||||||
|
}
|
||||||
|
}
|
||||||
167
resources/sillytavern/presets/role_play_less_impersonation.json
Normal file
@@ -0,0 +1,167 @@
|
|||||||
|
{
|
||||||
|
"instruct": {
|
||||||
|
"input_sequence": "<|start_header_id|>writer character {{user}}<|end_header_id|>\n\n",
|
||||||
|
"output_sequence": "<|start_header_id|>writer character {{char}}<|end_header_id|>\n\n",
|
||||||
|
"first_output_sequence": "",
|
||||||
|
"last_output_sequence": "<|start_header_id|>user<|end_header_id|>\n\nThe next message is from {{char}}.<|eot_id|><|start_header_id|>writer character {{char}}<|end_header_id|>\n\n",
|
||||||
|
"system_sequence_prefix": "",
|
||||||
|
"system_sequence_suffix": "",
|
||||||
|
"stop_sequence": "<|eot_id|>",
|
||||||
|
"wrap": false,
|
||||||
|
"macro": true,
|
||||||
|
"activation_regex": "",
|
||||||
|
"skip_examples": false,
|
||||||
|
"output_suffix": "<|eot_id|>",
|
||||||
|
"input_suffix": "<|eot_id|>",
|
||||||
|
"system_sequence": "<|start_header_id|>user<|end_header_id|>\n\n",
|
||||||
|
"system_suffix": "<|eot_id|>",
|
||||||
|
"user_alignment_message": "",
|
||||||
|
"last_system_sequence": "",
|
||||||
|
"system_same_as_user": false,
|
||||||
|
"names_behavior": "none",
|
||||||
|
"first_input_sequence": "",
|
||||||
|
"last_input_sequence": "",
|
||||||
|
"names_force_groups": true,
|
||||||
|
"name": "DreamGen Role-Play Lucid V1 Llama3"
|
||||||
|
},
|
||||||
|
"context": {
|
||||||
|
"story_string": "<|start_header_id|>system<|end_header_id|>\n\nYou are a skilled role-play writer and writing assistant with expertise across all genres.\n\nYou will perform several tasks, switching roles as needed:\n\n- Role-Playing: Use the `writer` role to write a role-play based on the provided information (plot, style, setting, characters, etc.) and user instructions. Use the `writer character <character_name>` role for dialog or when acting as a specific character, use the `writer narrative` role for narration.\n- Other: Use the `assistant` role for any other tasks the user may request.\n\n# Role-Play Information\n\n## Plot\n\n{{#if scenario}}{{scenario}}{{else}}Conversation between {{char}} and {{user}}.{{/if}}{{#if wiBefore}}\n\n## Setting\n\n{{wiBefore}}{{#if wiAfter}}\n\n{{wiAfter}}{{/if}}{{else}}{{#if wiAfter}}\n\n## Setting\n\n{{wiAfter}}{{/if}}{{/if}}\n\n## Characters\n\n### {{user}}\n\n{{#if persona}}{{persona}}{{else}}{{user}} is the protagonist of the role-play.{{/if}}\n\n### {{char}}\n\n{{#if description}}{{description}}\n\n{{/if}}{{#if personality}}{{personality}}\n\n{{/if}}<|eot_id|>",
|
||||||
|
"example_separator": "",
|
||||||
|
"chat_start": "",
|
||||||
|
"use_stop_strings": false,
|
||||||
|
"allow_jailbreak": false,
|
||||||
|
"names_as_stop_strings": true,
|
||||||
|
"always_force_name2": false,
|
||||||
|
"trim_sentences": false,
|
||||||
|
"single_line": false,
|
||||||
|
"name": "DreamGen Role-Play Lucid V1 Llama3"
|
||||||
|
},
|
||||||
|
"sysprompt": {
|
||||||
|
"name": "Blank",
|
||||||
|
"content": ""
|
||||||
|
},
|
||||||
|
"preset": {
|
||||||
|
"temp": 0.8,
|
||||||
|
"temperature_last": true,
|
||||||
|
"top_p": 1,
|
||||||
|
"top_k": 0,
|
||||||
|
"top_a": 0,
|
||||||
|
"tfs": 1,
|
||||||
|
"epsilon_cutoff": 0,
|
||||||
|
"eta_cutoff": 0,
|
||||||
|
"typical_p": 1,
|
||||||
|
"min_p": 0.05,
|
||||||
|
"rep_pen": 1,
|
||||||
|
"rep_pen_range": 0,
|
||||||
|
"rep_pen_decay": 0,
|
||||||
|
"rep_pen_slope": 1,
|
||||||
|
"no_repeat_ngram_size": 0,
|
||||||
|
"penalty_alpha": 0,
|
||||||
|
"num_beams": 1,
|
||||||
|
"length_penalty": 1,
|
||||||
|
"min_length": 0,
|
||||||
|
"encoder_rep_pen": 1,
|
||||||
|
"freq_pen": 0,
|
||||||
|
"presence_pen": 0,
|
||||||
|
"skew": 0,
|
||||||
|
"do_sample": true,
|
||||||
|
"early_stopping": false,
|
||||||
|
"dynatemp": false,
|
||||||
|
"min_temp": 0,
|
||||||
|
"max_temp": 2,
|
||||||
|
"dynatemp_exponent": 1,
|
||||||
|
"smoothing_factor": 0,
|
||||||
|
"smoothing_curve": 1,
|
||||||
|
"dry_allowed_length": 2,
|
||||||
|
"dry_multiplier": 0.8,
|
||||||
|
"dry_base": 1.75,
|
||||||
|
"dry_sequence_breakers": "[\"\\n\", \":\", \"\\\"\", \"*\"]",
|
||||||
|
"dry_penalty_last_n": 0,
|
||||||
|
"add_bos_token": true,
|
||||||
|
"ban_eos_token": false,
|
||||||
|
"skip_special_tokens": false,
|
||||||
|
"mirostat_mode": 0,
|
||||||
|
"mirostat_tau": 5,
|
||||||
|
"mirostat_eta": 0.1,
|
||||||
|
"guidance_scale": 1,
|
||||||
|
"negative_prompt": "",
|
||||||
|
"grammar_string": "",
|
||||||
|
"json_schema": {},
|
||||||
|
"banned_tokens": "",
|
||||||
|
"sampler_priority": [
|
||||||
|
"repetition_penalty",
|
||||||
|
"presence_penalty",
|
||||||
|
"frequency_penalty",
|
||||||
|
"dry",
|
||||||
|
"temperature",
|
||||||
|
"dynamic_temperature",
|
||||||
|
"quadratic_sampling",
|
||||||
|
"top_k",
|
||||||
|
"top_p",
|
||||||
|
"typical_p",
|
||||||
|
"epsilon_cutoff",
|
||||||
|
"eta_cutoff",
|
||||||
|
"tfs",
|
||||||
|
"top_a",
|
||||||
|
"min_p",
|
||||||
|
"mirostat",
|
||||||
|
"xtc",
|
||||||
|
"encoder_repetition_penalty",
|
||||||
|
"no_repeat_ngram"
|
||||||
|
],
|
||||||
|
"samplers": [
|
||||||
|
"dry",
|
||||||
|
"top_k",
|
||||||
|
"tfs_z",
|
||||||
|
"typical_p",
|
||||||
|
"top_p",
|
||||||
|
"min_p",
|
||||||
|
"xtc",
|
||||||
|
"temperature"
|
||||||
|
],
|
||||||
|
"samplers_priorities": [
|
||||||
|
"dry",
|
||||||
|
"penalties",
|
||||||
|
"no_repeat_ngram",
|
||||||
|
"temperature",
|
||||||
|
"top_nsigma",
|
||||||
|
"top_p_top_k",
|
||||||
|
"top_a",
|
||||||
|
"min_p",
|
||||||
|
"tfs",
|
||||||
|
"eta_cutoff",
|
||||||
|
"epsilon_cutoff",
|
||||||
|
"typical_p",
|
||||||
|
"quadratic",
|
||||||
|
"xtc"
|
||||||
|
],
|
||||||
|
"ignore_eos_token": false,
|
||||||
|
"spaces_between_special_tokens": true,
|
||||||
|
"speculative_ngram": false,
|
||||||
|
"sampler_order": [
|
||||||
|
6,
|
||||||
|
0,
|
||||||
|
1,
|
||||||
|
3,
|
||||||
|
4,
|
||||||
|
2,
|
||||||
|
5
|
||||||
|
],
|
||||||
|
"logit_bias": [],
|
||||||
|
"xtc_threshold": 0.1,
|
||||||
|
"xtc_probability": 0,
|
||||||
|
"nsigma": 0,
|
||||||
|
"ignore_eos_token_aphrodite": false,
|
||||||
|
"spaces_between_special_tokens_aphrodite": true,
|
||||||
|
"rep_pen_size": 0,
|
||||||
|
"genamt": 1000,
|
||||||
|
"max_length": 32256,
|
||||||
|
"name": "Lucid V1 Nemo"
|
||||||
|
},
|
||||||
|
"reasoning": {
|
||||||
|
"name": "DreamGen Lucid V1",
|
||||||
|
"prefix": "<|reasoning_start|>",
|
||||||
|
"suffix": "<|reasoning_end|>",
|
||||||
|
"separator": ""
|
||||||
|
}
|
||||||
|
}
|
||||||
167
resources/sillytavern/presets/writing_basic.json
Normal file
@@ -0,0 +1,167 @@
|
|||||||
|
{
|
||||||
|
"instruct": {
|
||||||
|
"input_sequence": "<|start_header_id|>writer<|end_header_id|>\n\n",
|
||||||
|
"output_sequence": "<|start_header_id|>writer<|end_header_id|>\n\n",
|
||||||
|
"first_output_sequence": "",
|
||||||
|
"last_output_sequence": "",
|
||||||
|
"system_sequence_prefix": "",
|
||||||
|
"system_sequence_suffix": "",
|
||||||
|
"stop_sequence": "",
|
||||||
|
"wrap": false,
|
||||||
|
"macro": true,
|
||||||
|
"activation_regex": "",
|
||||||
|
"skip_examples": false,
|
||||||
|
"output_suffix": "<|eot_id|>",
|
||||||
|
"input_suffix": "<|eot_id|>",
|
||||||
|
"system_sequence": "<|start_header_id|>user<|end_header_id|>\n\n",
|
||||||
|
"system_suffix": "<|eot_id|>",
|
||||||
|
"user_alignment_message": "",
|
||||||
|
"last_system_sequence": "",
|
||||||
|
"system_same_as_user": false,
|
||||||
|
"names_behavior": "none",
|
||||||
|
"first_input_sequence": "",
|
||||||
|
"last_input_sequence": "",
|
||||||
|
"names_force_groups": true,
|
||||||
|
"name": "DreamGen Writing Lucid V1 Llama3"
|
||||||
|
},
|
||||||
|
"context": {
|
||||||
|
"story_string": "<|start_header_id|>system<|end_header_id|>\n\nYou are a skilled fiction writer and writing assistant with expertise across all genres.\n\nYou will perform several tasks, switching roles as needed:\n\n- Writing: Use the `writer` role to write a role-play based on the provided information (plot, style, setting, characters, etc.) and user instructions.\n- Other: Use the `assistant` role for any other tasks the user may request.\n\n# Story Information\n\n## Plot\n\n{{#if scenario}}{{scenario}}{{else}}Conversation between {{char}} and {{user}}.{{/if}}{{#if wiBefore}}\n\n## Setting\n\n{{wiBefore}}{{#if wiAfter}}\n\n{{wiAfter}}{{/if}}{{else}}{{#if wiAfter}}\n\n## Setting\n\n{{wiAfter}}{{/if}}{{/if}}\n\n{{#if description}}{{description}}\n\n{{/if}}<|eot_id|>",
|
||||||
|
"example_separator": "",
|
||||||
|
"chat_start": "",
|
||||||
|
"use_stop_strings": false,
|
||||||
|
"allow_jailbreak": false,
|
||||||
|
"names_as_stop_strings": true,
|
||||||
|
"always_force_name2": false,
|
||||||
|
"trim_sentences": false,
|
||||||
|
"single_line": false,
|
||||||
|
"name": "DreamGen Writing Lucid V1 Llama3"
|
||||||
|
},
|
||||||
|
"sysprompt": {
|
||||||
|
"name": "Blank",
|
||||||
|
"content": ""
|
||||||
|
},
|
||||||
|
"preset": {
|
||||||
|
"temp": 0.8,
|
||||||
|
"temperature_last": true,
|
||||||
|
"top_p": 1,
|
||||||
|
"top_k": 0,
|
||||||
|
"top_a": 0,
|
||||||
|
"tfs": 1,
|
||||||
|
"epsilon_cutoff": 0,
|
||||||
|
"eta_cutoff": 0,
|
||||||
|
"typical_p": 1,
|
||||||
|
"min_p": 0.05,
|
||||||
|
"rep_pen": 1,
|
||||||
|
"rep_pen_range": 0,
|
||||||
|
"rep_pen_decay": 0,
|
||||||
|
"rep_pen_slope": 1,
|
||||||
|
"no_repeat_ngram_size": 0,
|
||||||
|
"penalty_alpha": 0,
|
||||||
|
"num_beams": 1,
|
||||||
|
"length_penalty": 1,
|
||||||
|
"min_length": 0,
|
||||||
|
"encoder_rep_pen": 1,
|
||||||
|
"freq_pen": 0,
|
||||||
|
"presence_pen": 0,
|
||||||
|
"skew": 0,
|
||||||
|
"do_sample": true,
|
||||||
|
"early_stopping": false,
|
||||||
|
"dynatemp": false,
|
||||||
|
"min_temp": 0,
|
||||||
|
"max_temp": 2,
|
||||||
|
"dynatemp_exponent": 1,
|
||||||
|
"smoothing_factor": 0,
|
||||||
|
"smoothing_curve": 1,
|
||||||
|
"dry_allowed_length": 2,
|
||||||
|
"dry_multiplier": 0.8,
|
||||||
|
"dry_base": 1.75,
|
||||||
|
"dry_sequence_breakers": "[\"\\n\", \":\", \"\\\"\", \"*\"]",
|
||||||
|
"dry_penalty_last_n": 0,
|
||||||
|
"add_bos_token": true,
|
||||||
|
"ban_eos_token": false,
|
||||||
|
"skip_special_tokens": false,
|
||||||
|
"mirostat_mode": 0,
|
||||||
|
"mirostat_tau": 5,
|
||||||
|
"mirostat_eta": 0.1,
|
||||||
|
"guidance_scale": 1,
|
||||||
|
"negative_prompt": "",
|
||||||
|
"grammar_string": "",
|
||||||
|
"json_schema": {},
|
||||||
|
"banned_tokens": "",
|
||||||
|
"sampler_priority": [
|
||||||
|
"repetition_penalty",
|
||||||
|
"presence_penalty",
|
||||||
|
"frequency_penalty",
|
||||||
|
"dry",
|
||||||
|
"temperature",
|
||||||
|
"dynamic_temperature",
|
||||||
|
"quadratic_sampling",
|
||||||
|
"top_k",
|
||||||
|
"top_p",
|
||||||
|
"typical_p",
|
||||||
|
"epsilon_cutoff",
|
||||||
|
"eta_cutoff",
|
||||||
|
"tfs",
|
||||||
|
"top_a",
|
||||||
|
"min_p",
|
||||||
|
"mirostat",
|
||||||
|
"xtc",
|
||||||
|
"encoder_repetition_penalty",
|
||||||
|
"no_repeat_ngram"
|
||||||
|
],
|
||||||
|
"samplers": [
|
||||||
|
"dry",
|
||||||
|
"top_k",
|
||||||
|
"tfs_z",
|
||||||
|
"typical_p",
|
||||||
|
"top_p",
|
||||||
|
"min_p",
|
||||||
|
"xtc",
|
||||||
|
"temperature"
|
||||||
|
],
|
||||||
|
"samplers_priorities": [
|
||||||
|
"dry",
|
||||||
|
"penalties",
|
||||||
|
"no_repeat_ngram",
|
||||||
|
"temperature",
|
||||||
|
"top_nsigma",
|
||||||
|
"top_p_top_k",
|
||||||
|
"top_a",
|
||||||
|
"min_p",
|
||||||
|
"tfs",
|
||||||
|
"eta_cutoff",
|
||||||
|
"epsilon_cutoff",
|
||||||
|
"typical_p",
|
||||||
|
"quadratic",
|
||||||
|
"xtc"
|
||||||
|
],
|
||||||
|
"ignore_eos_token": false,
|
||||||
|
"spaces_between_special_tokens": true,
|
||||||
|
"speculative_ngram": false,
|
||||||
|
"sampler_order": [
|
||||||
|
6,
|
||||||
|
0,
|
||||||
|
1,
|
||||||
|
3,
|
||||||
|
4,
|
||||||
|
2,
|
||||||
|
5
|
||||||
|
],
|
||||||
|
"logit_bias": [],
|
||||||
|
"xtc_threshold": 0.1,
|
||||||
|
"xtc_probability": 0,
|
||||||
|
"nsigma": 0,
|
||||||
|
"ignore_eos_token_aphrodite": false,
|
||||||
|
"spaces_between_special_tokens_aphrodite": true,
|
||||||
|
"rep_pen_size": 0,
|
||||||
|
"genamt": 250,
|
||||||
|
"max_length": 32256,
|
||||||
|
"name": "Lucid V1 Nemo"
|
||||||
|
},
|
||||||
|
"reasoning": {
|
||||||
|
"name": "DreamGen Lucid V1",
|
||||||
|
"prefix": "<|reasoning_start|>",
|
||||||
|
"suffix": "<|reasoning_end|>",
|
||||||
|
"separator": ""
|
||||||
|
}
|
||||||
|
}
|
||||||
37
special_tokens_map.json
Normal file
@@ -0,0 +1,37 @@
|
|||||||
|
{
|
||||||
|
"additional_special_tokens": [
|
||||||
|
"<|start_header_id|>",
|
||||||
|
"<|end_header_id|>",
|
||||||
|
"<|eot_id|>",
|
||||||
|
"<|reasoning_start|>",
|
||||||
|
"<|reasoning_end|>"
|
||||||
|
],
|
||||||
|
"bos_token": {
|
||||||
|
"content": "<s>",
|
||||||
|
"lstrip": false,
|
||||||
|
"normalized": false,
|
||||||
|
"rstrip": false,
|
||||||
|
"single_word": false
|
||||||
|
},
|
||||||
|
"eos_token": {
|
||||||
|
"content": "</s>",
|
||||||
|
"lstrip": false,
|
||||||
|
"normalized": false,
|
||||||
|
"rstrip": false,
|
||||||
|
"single_word": false
|
||||||
|
},
|
||||||
|
"pad_token": {
|
||||||
|
"content": "<pad>",
|
||||||
|
"lstrip": false,
|
||||||
|
"normalized": false,
|
||||||
|
"rstrip": false,
|
||||||
|
"single_word": false
|
||||||
|
},
|
||||||
|
"unk_token": {
|
||||||
|
"content": "<unk>",
|
||||||
|
"lstrip": false,
|
||||||
|
"normalized": false,
|
||||||
|
"rstrip": false,
|
||||||
|
"single_word": false
|
||||||
|
}
|
||||||
|
}
|
||||||
3
tokenizer.json
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:00670795773e235e174ca35f6e9b3dc3ea3f4eff6e2f03757132b561b55c65a0
|
||||||
|
size 17079259
|
||||||