初始化项目,由ModelHub XC社区提供模型

Model: peiyi9979/math-shepherd-mistral-7b-prm
Source: Original Platform
This commit is contained in:
ModelHub XC
2026-04-09 14:34:24 +08:00
commit c71bac85dc
9 changed files with 467 additions and 0 deletions

35
.gitattributes vendored Normal file
View File

@@ -0,0 +1,35 @@
*.7z filter=lfs diff=lfs merge=lfs -text
*.arrow filter=lfs diff=lfs merge=lfs -text
*.bin filter=lfs diff=lfs merge=lfs -text
*.bz2 filter=lfs diff=lfs merge=lfs -text
*.ckpt filter=lfs diff=lfs merge=lfs -text
*.ftz filter=lfs diff=lfs merge=lfs -text
*.gz filter=lfs diff=lfs merge=lfs -text
*.h5 filter=lfs diff=lfs merge=lfs -text
*.joblib filter=lfs diff=lfs merge=lfs -text
*.lfs.* filter=lfs diff=lfs merge=lfs -text
*.mlmodel filter=lfs diff=lfs merge=lfs -text
*.model filter=lfs diff=lfs merge=lfs -text
*.msgpack filter=lfs diff=lfs merge=lfs -text
*.npy filter=lfs diff=lfs merge=lfs -text
*.npz filter=lfs diff=lfs merge=lfs -text
*.onnx filter=lfs diff=lfs merge=lfs -text
*.ot filter=lfs diff=lfs merge=lfs -text
*.parquet filter=lfs diff=lfs merge=lfs -text
*.pb filter=lfs diff=lfs merge=lfs -text
*.pickle filter=lfs diff=lfs merge=lfs -text
*.pkl filter=lfs diff=lfs merge=lfs -text
*.pt filter=lfs diff=lfs merge=lfs -text
*.pth filter=lfs diff=lfs merge=lfs -text
*.rar filter=lfs diff=lfs merge=lfs -text
*.safetensors filter=lfs diff=lfs merge=lfs -text
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
*.tar.* filter=lfs diff=lfs merge=lfs -text
*.tar filter=lfs diff=lfs merge=lfs -text
*.tflite filter=lfs diff=lfs merge=lfs -text
*.tgz filter=lfs diff=lfs merge=lfs -text
*.wasm filter=lfs diff=lfs merge=lfs -text
*.xz filter=lfs diff=lfs merge=lfs -text
*.zip filter=lfs diff=lfs merge=lfs -text
*.zst filter=lfs diff=lfs merge=lfs -text
*tfevents* filter=lfs diff=lfs merge=lfs -text

41
README.md Normal file
View File

@@ -0,0 +1,41 @@
Process reward model (mistral-7b) used in [Math-Shepherd](https://arxiv.org/pdf/2312.08935.pdf).
`Input`: question + step-by-step solutions with a special step tag `ки`, e.g.,
```
Janet\u2019s ducks lay 16 eggs per day. She eats three for breakfast every morning and bakes .... ? Step 1: Janet's ducks lay 16 eggs per day. ки\nStep 2: She eats three for breakfast every morning, so she has 16 - 3 = 13 eggs left. ки\nStep 3: She bakes muffins for her friends every day with four eggs, so she has 13 - 4 = 9 eggs left. ки\nStep 4: She sells the remainder at the farmers' market daily for $2 per fresh duck egg, so she makes 9 * $2 = $18 every day at the farmers' market. The answer is: 18 ки
```
`Output`: the logits. You need to post-process it to achieve the score of each step.
```python
from transformers import AutoTokenizer
from transformers import AutoModelForCausalLM
import torch
good_token = '+'
bad_token = '-'
step_tag = 'ки'
tokenizer = AutoTokenizer.from_pretrained('peiyi9979/math-shepherd-mistral-7b-prm')
candidate_tokens = tokenizer.encode(f"{good_token} {bad_token}")[1:] # [648, 387]
step_tag_id = tokenizer.encode(f"{step_tag}")[-1] # 12902
model = AutoModelForCausalLM.from_pretrained('peiyi9979/math-shepherd-mistral-7b-prm').eval()
question = """Janet\u2019s ducks lay 16 eggs per day. She eats three for breakfast every morning and bakes muffins for her friends every day with four. She sells the remainder at the farmers' market daily for $2 per fresh duck egg. How much in dollars does she make every day at the farmers' market?"""
output1 = """Step 1: Janet's ducks lay 16 eggs per day. ки\nStep 2: She eats three for breakfast every morning, so she has 16 - 3 = 13 eggs left. ки\nStep 3: She bakes muffins for her friends every day with four eggs, so she has 13 - 4 = 9 eggs left. ки\nStep 4: She sells the remainder at the farmers' market daily for $2 per fresh duck egg, so she makes 9 * $2 = $18 every day at the farmers' market. The answer is: 18 ки""" # 18 is right
output2 = """Step 1: Janet's ducks lay 16 eggs per day. ки\nStep 2: She eats three for breakfast every morning, so she has 16 - 3 = 13 eggs left. ки\nStep 3: She bakes muffins for her friends every day with four eggs, so she has 13 - 4 = 9 eggs left. ки\nStep 4: She sells the remainder at the farmers' market daily for $2 per fresh duck egg, so she makes 9 * $2 = $17 every day at the farmers' market. The answer is: 17 ки""" # 17 is wrong
for output in [output1, output2]:
input_for_prm = f"{question} {output}"
input_id = torch.tensor([tokenizer.encode(input_for_prm)])
with torch.no_grad():
logits = model(input_id).logits[:,:,candidate_tokens]
scores = logits.softmax(dim=-1)[:,:,0]
step_scores = scores[input_id == step_tag_id]
print(step_scores)
# tensor([0.9955, 0.9958, 0.9983, 0.9957])
# tensor([0.9955, 0.9958, 0.9983, 0.0240])
```

25
config.json Normal file
View File

@@ -0,0 +1,25 @@
{
"architectures": [
"LlamaForCausalLM"
],
"bos_token_id": 1,
"eos_token_id": 2,
"hidden_act": "silu",
"hidden_size": 4096,
"initializer_range": 0.02,
"intermediate_size": 14336,
"max_position_embeddings": 4096,
"model_type": "llama",
"num_attention_heads": 32,
"num_hidden_layers": 32,
"num_key_value_heads": 8,
"pretraining_tp": 1,
"rms_norm_eps": 1e-05,
"rope_scaling": null,
"rope_theta": 10000.0,
"tie_word_embeddings": false,
"torch_dtype": "bfloat16",
"transformers_version": "4.33.1",
"use_cache": true,
"vocab_size": 32000
}

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:7942123dacfadc7bd7aeccdb19a521f69c326ab07e73376489e84c5abcbb3d75
size 9492067331

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:dec791d4d9bcd8ba8fec1fe71951041c9aa7316a043e6db47f40368724313855
size 6333666262

View File

@@ -0,0 +1,298 @@
{
"metadata": {
"total_size": 14483464192
},
"weight_map": {
"model.embed_tokens.weight": "pytorch_model-00001-of-000002.bin",
"model.norm.weight": "pytorch_model-00001-of-000002.bin",
"lm_head.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.0.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.0.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.0.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.0.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.0.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.0.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.0.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.0.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.0.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.1.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.1.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.1.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.1.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.1.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.1.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.1.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.1.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.1.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.2.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.2.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.2.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.2.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.2.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.2.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.2.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.2.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.2.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.3.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.3.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.3.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.3.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.3.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.3.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.3.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.3.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.3.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.4.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.4.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.4.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.4.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.4.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.4.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.4.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.4.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.4.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.5.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.5.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.5.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.5.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.5.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.5.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.5.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.5.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.5.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.6.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.6.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.6.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.6.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.6.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.6.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.6.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.6.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.6.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.7.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.7.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.7.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.7.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.7.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.7.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.7.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.7.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.7.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.8.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.8.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.8.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.8.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.8.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.8.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.8.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.8.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.8.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.9.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.9.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.9.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.9.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.9.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.9.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.9.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.9.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.9.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.10.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.10.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.10.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.10.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.10.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.10.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.10.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.10.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.10.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.11.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.11.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.11.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.11.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.11.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.11.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.11.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.11.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.11.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.12.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.12.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.12.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.12.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.12.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.12.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.12.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.12.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.12.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.13.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.13.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.13.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.13.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.13.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.13.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.13.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.13.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.13.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.14.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.14.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.14.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.14.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.14.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.14.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.14.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.14.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.14.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.15.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.15.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.15.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.15.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.15.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.15.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.15.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.15.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.15.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.16.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.16.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.16.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.16.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.16.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.16.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.16.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.16.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.16.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.17.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.17.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.17.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.17.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.17.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.17.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.17.mlp.down_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.17.input_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.17.post_attention_layernorm.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.18.self_attn.q_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.18.self_attn.k_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.18.self_attn.v_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.18.self_attn.o_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.18.mlp.gate_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.18.mlp.up_proj.weight": "pytorch_model-00001-of-000002.bin",
"model.layers.18.mlp.down_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.18.input_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.18.post_attention_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.19.self_attn.q_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.19.self_attn.k_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.19.self_attn.v_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.19.self_attn.o_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.19.mlp.gate_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.19.mlp.up_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.19.mlp.down_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.19.input_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.19.post_attention_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.20.self_attn.q_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.20.self_attn.k_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.20.self_attn.v_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.20.self_attn.o_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.20.mlp.gate_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.20.mlp.up_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.20.mlp.down_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.20.input_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.20.post_attention_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.21.self_attn.q_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.21.self_attn.k_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.21.self_attn.v_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.21.self_attn.o_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.21.mlp.gate_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.21.mlp.up_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.21.mlp.down_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.21.input_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.21.post_attention_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.22.self_attn.q_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.22.self_attn.k_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.22.self_attn.v_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.22.self_attn.o_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.22.mlp.gate_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.22.mlp.up_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.22.mlp.down_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.22.input_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.22.post_attention_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.23.self_attn.q_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.23.self_attn.k_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.23.self_attn.v_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.23.self_attn.o_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.23.mlp.gate_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.23.mlp.up_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.23.mlp.down_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.23.input_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.23.post_attention_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.24.self_attn.q_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.24.self_attn.k_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.24.self_attn.v_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.24.self_attn.o_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.24.mlp.gate_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.24.mlp.up_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.24.mlp.down_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.24.input_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.24.post_attention_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.25.self_attn.q_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.25.self_attn.k_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.25.self_attn.v_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.25.self_attn.o_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.25.mlp.gate_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.25.mlp.up_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.25.mlp.down_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.25.input_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.25.post_attention_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.26.self_attn.q_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.26.self_attn.k_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.26.self_attn.v_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.26.self_attn.o_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.26.mlp.gate_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.26.mlp.up_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.26.mlp.down_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.26.input_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.26.post_attention_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.27.self_attn.q_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.27.self_attn.k_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.27.self_attn.v_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.27.self_attn.o_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.27.mlp.gate_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.27.mlp.up_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.27.mlp.down_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.27.input_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.27.post_attention_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.28.self_attn.q_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.28.self_attn.k_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.28.self_attn.v_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.28.self_attn.o_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.28.mlp.gate_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.28.mlp.up_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.28.mlp.down_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.28.input_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.28.post_attention_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.29.self_attn.q_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.29.self_attn.k_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.29.self_attn.v_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.29.self_attn.o_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.29.mlp.gate_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.29.mlp.up_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.29.mlp.down_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.29.input_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.29.post_attention_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.30.self_attn.q_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.30.self_attn.k_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.30.self_attn.v_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.30.self_attn.o_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.30.mlp.gate_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.30.mlp.up_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.30.mlp.down_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.30.input_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.30.post_attention_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.31.self_attn.q_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.31.self_attn.k_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.31.self_attn.v_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.31.self_attn.o_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.31.mlp.gate_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.31.mlp.up_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.31.mlp.down_proj.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.31.input_layernorm.weight": "pytorch_model-00002-of-000002.bin",
"model.layers.31.post_attention_layernorm.weight": "pytorch_model-00002-of-000002.bin"
}
}

23
special_tokens_map.json Normal file
View File

@@ -0,0 +1,23 @@
{
"bos_token": {
"content": "<s>",
"lstrip": false,
"normalized": true,
"rstrip": false,
"single_word": false
},
"eos_token": {
"content": "</s>",
"lstrip": false,
"normalized": true,
"rstrip": false,
"single_word": false
},
"unk_token": {
"content": "<unk>",
"lstrip": false,
"normalized": true,
"rstrip": false,
"single_word": false
}
}

BIN
tokenizer.model (Stored with Git LFS) Normal file

Binary file not shown.

36
tokenizer_config.json Normal file
View File

@@ -0,0 +1,36 @@
{
"add_bos_token": true,
"add_eos_token": false,
"bos_token": {
"__type": "AddedToken",
"content": "<s>",
"lstrip": false,
"normalized": true,
"rstrip": false,
"single_word": false
},
"clean_up_tokenization_spaces": false,
"eos_token": {
"__type": "AddedToken",
"content": "</s>",
"lstrip": false,
"normalized": true,
"rstrip": false,
"single_word": false
},
"legacy": null,
"model_max_length": 1000000000000000019884624838656,
"pad_token": null,
"sp_model_kwargs": {},
"spaces_between_special_tokens": false,
"tokenizer_class": "LlamaTokenizer",
"unk_token": {
"__type": "AddedToken",
"content": "<unk>",
"lstrip": false,
"normalized": true,
"rstrip": false,
"single_word": false
},
"use_default_system_prompt": true
}