初始化项目,由ModelHub XC社区提供模型
Model: jackf857/llama-3-8b-base-ipo-ultrafeedback-4xh200-batch-128-20260428-004616 Source: Original Platform
This commit is contained in:
36
.gitattributes
vendored
Normal file
36
.gitattributes
vendored
Normal file
@@ -0,0 +1,36 @@
|
||||
*.7z filter=lfs diff=lfs merge=lfs -text
|
||||
*.arrow filter=lfs diff=lfs merge=lfs -text
|
||||
*.bin filter=lfs diff=lfs merge=lfs -text
|
||||
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
||||
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
||||
*.ftz filter=lfs diff=lfs merge=lfs -text
|
||||
*.gz filter=lfs diff=lfs merge=lfs -text
|
||||
*.h5 filter=lfs diff=lfs merge=lfs -text
|
||||
*.joblib filter=lfs diff=lfs merge=lfs -text
|
||||
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
||||
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
||||
*.model filter=lfs diff=lfs merge=lfs -text
|
||||
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
||||
*.npy filter=lfs diff=lfs merge=lfs -text
|
||||
*.npz filter=lfs diff=lfs merge=lfs -text
|
||||
*.onnx filter=lfs diff=lfs merge=lfs -text
|
||||
*.ot filter=lfs diff=lfs merge=lfs -text
|
||||
*.parquet filter=lfs diff=lfs merge=lfs -text
|
||||
*.pb filter=lfs diff=lfs merge=lfs -text
|
||||
*.pickle filter=lfs diff=lfs merge=lfs -text
|
||||
*.pkl filter=lfs diff=lfs merge=lfs -text
|
||||
*.pt filter=lfs diff=lfs merge=lfs -text
|
||||
*.pth filter=lfs diff=lfs merge=lfs -text
|
||||
*.rar filter=lfs diff=lfs merge=lfs -text
|
||||
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
||||
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
||||
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
||||
*.tar filter=lfs diff=lfs merge=lfs -text
|
||||
*.tflite filter=lfs diff=lfs merge=lfs -text
|
||||
*.tgz filter=lfs diff=lfs merge=lfs -text
|
||||
*.wasm filter=lfs diff=lfs merge=lfs -text
|
||||
*.xz filter=lfs diff=lfs merge=lfs -text
|
||||
*.zip filter=lfs diff=lfs merge=lfs -text
|
||||
*.zst filter=lfs diff=lfs merge=lfs -text
|
||||
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
||||
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
||||
76
README.md
Normal file
76
README.md
Normal file
@@ -0,0 +1,76 @@
|
||||
---
|
||||
library_name: transformers
|
||||
base_model: W-61/llama-3-8b-base-sft-ultrachat-8xh200
|
||||
tags:
|
||||
- alignment-handbook
|
||||
- ipo
|
||||
- generated_from_trainer
|
||||
datasets:
|
||||
- HuggingFaceH4/ultrafeedback_binarized
|
||||
model-index:
|
||||
- name: llama-3-8b-base-ipo-ultrafeedback-4xh200-batch-128-20260428-004616
|
||||
results: []
|
||||
---
|
||||
|
||||
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
||||
should probably proofread and complete it, then remove this comment. -->
|
||||
|
||||
# llama-3-8b-base-ipo-ultrafeedback-4xh200-batch-128-20260428-004616
|
||||
|
||||
This model is a fine-tuned version of [W-61/llama-3-8b-base-sft-ultrachat-8xh200](https://huggingface.co/W-61/llama-3-8b-base-sft-ultrachat-8xh200) on the HuggingFaceH4/ultrafeedback_binarized dataset.
|
||||
It achieves the following results on the evaluation set:
|
||||
- Loss: 2313.8057
|
||||
- Rewards/chosen: -0.0439
|
||||
- Rewards/rejected: -0.0689
|
||||
- Rewards/accuracies: 0.6800
|
||||
- Rewards/margins: 0.0250
|
||||
- Logps/rejected: -8.1847
|
||||
- Logps/chosen: -5.5025
|
||||
- Logits/rejected: -0.2777
|
||||
- Logits/chosen: -0.2620
|
||||
|
||||
## Model description
|
||||
|
||||
More information needed
|
||||
|
||||
## Intended uses & limitations
|
||||
|
||||
More information needed
|
||||
|
||||
## Training and evaluation data
|
||||
|
||||
More information needed
|
||||
|
||||
## Training procedure
|
||||
|
||||
### Training hyperparameters
|
||||
|
||||
The following hyperparameters were used during training:
|
||||
- learning_rate: 5e-07
|
||||
- train_batch_size: 4
|
||||
- eval_batch_size: 4
|
||||
- seed: 42
|
||||
- distributed_type: multi-GPU
|
||||
- num_devices: 4
|
||||
- gradient_accumulation_steps: 8
|
||||
- total_train_batch_size: 128
|
||||
- total_eval_batch_size: 16
|
||||
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
||||
- lr_scheduler_type: cosine
|
||||
- lr_scheduler_warmup_ratio: 0.1
|
||||
- num_epochs: 1
|
||||
|
||||
### Training results
|
||||
|
||||
| Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
|
||||
|:-------------:|:------:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
|
||||
| 19275.1328 | 0.4188 | 200 | 2417.4961 | -0.0134 | -0.0217 | 0.6560 | 0.0083 | -3.4695 | -2.4551 | -0.6300 | -0.6352 |
|
||||
| 18486.2438 | 0.8377 | 400 | 2313.8057 | -0.0439 | -0.0689 | 0.6800 | 0.0250 | -8.1847 | -5.5025 | -0.2777 | -0.2620 |
|
||||
|
||||
|
||||
### Framework versions
|
||||
|
||||
- Transformers 4.51.0
|
||||
- Pytorch 2.3.1+cu121
|
||||
- Datasets 2.21.0
|
||||
- Tokenizers 0.21.4
|
||||
22
all_results.json
Normal file
22
all_results.json
Normal file
@@ -0,0 +1,22 @@
|
||||
{
|
||||
"epoch": 0.9989528795811519,
|
||||
"eval_logits/chosen": -0.1925041377544403,
|
||||
"eval_logits/rejected": -0.21385852992534637,
|
||||
"eval_logps/chosen": -5.949731349945068,
|
||||
"eval_logps/rejected": -8.787737846374512,
|
||||
"eval_loss": 2305.42529296875,
|
||||
"eval_rewards/accuracies": 0.6759999990463257,
|
||||
"eval_rewards/chosen": -0.04833168908953667,
|
||||
"eval_rewards/margins": 0.026571419090032578,
|
||||
"eval_rewards/rejected": -0.07490310817956924,
|
||||
"eval_runtime": 78.3976,
|
||||
"eval_samples": 2000,
|
||||
"eval_samples_per_second": 25.511,
|
||||
"eval_steps_per_second": 1.594,
|
||||
"total_flos": 0.0,
|
||||
"train_loss": 19137.84001572327,
|
||||
"train_runtime": 6782.201,
|
||||
"train_samples": 61135,
|
||||
"train_samples_per_second": 9.014,
|
||||
"train_steps_per_second": 0.07
|
||||
}
|
||||
29
config.json
Normal file
29
config.json
Normal file
@@ -0,0 +1,29 @@
|
||||
{
|
||||
"architectures": [
|
||||
"LlamaForCausalLM"
|
||||
],
|
||||
"attention_bias": false,
|
||||
"attention_dropout": 0.0,
|
||||
"bos_token_id": 128000,
|
||||
"eos_token_id": 128001,
|
||||
"head_dim": 128,
|
||||
"hidden_act": "silu",
|
||||
"hidden_size": 4096,
|
||||
"initializer_range": 0.02,
|
||||
"intermediate_size": 14336,
|
||||
"max_position_embeddings": 8192,
|
||||
"mlp_bias": false,
|
||||
"model_type": "llama",
|
||||
"num_attention_heads": 32,
|
||||
"num_hidden_layers": 32,
|
||||
"num_key_value_heads": 8,
|
||||
"pretraining_tp": 1,
|
||||
"rms_norm_eps": 1e-05,
|
||||
"rope_scaling": null,
|
||||
"rope_theta": 500000.0,
|
||||
"tie_word_embeddings": false,
|
||||
"torch_dtype": "float32",
|
||||
"transformers_version": "4.51.0",
|
||||
"use_cache": true,
|
||||
"vocab_size": 128256
|
||||
}
|
||||
16
eval_results.json
Normal file
16
eval_results.json
Normal file
@@ -0,0 +1,16 @@
|
||||
{
|
||||
"epoch": 0.9989528795811519,
|
||||
"eval_logits/chosen": -0.1925041377544403,
|
||||
"eval_logits/rejected": -0.21385852992534637,
|
||||
"eval_logps/chosen": -5.949731349945068,
|
||||
"eval_logps/rejected": -8.787737846374512,
|
||||
"eval_loss": 2305.42529296875,
|
||||
"eval_rewards/accuracies": 0.6759999990463257,
|
||||
"eval_rewards/chosen": -0.04833168908953667,
|
||||
"eval_rewards/margins": 0.026571419090032578,
|
||||
"eval_rewards/rejected": -0.07490310817956924,
|
||||
"eval_runtime": 78.3976,
|
||||
"eval_samples": 2000,
|
||||
"eval_samples_per_second": 25.511,
|
||||
"eval_steps_per_second": 1.594
|
||||
}
|
||||
9
generation_config.json
Normal file
9
generation_config.json
Normal file
@@ -0,0 +1,9 @@
|
||||
{
|
||||
"bos_token_id": 128000,
|
||||
"do_sample": true,
|
||||
"eos_token_id": 128001,
|
||||
"max_length": 4096,
|
||||
"temperature": 0.6,
|
||||
"top_p": 0.9,
|
||||
"transformers_version": "4.51.0"
|
||||
}
|
||||
3
model-00001-of-00007.safetensors
Normal file
3
model-00001-of-00007.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:6d97f446094bc90c15d6a0eba9b9ce5ff1dda9991bb5a71605f45d8c187a354b
|
||||
size 4886466168
|
||||
3
model-00002-of-00007.safetensors
Normal file
3
model-00002-of-00007.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:52422dc4dcb6381718f555f360d1b84852eeb5eaba26f26133a445728662abee
|
||||
size 4832007448
|
||||
3
model-00003-of-00007.safetensors
Normal file
3
model-00003-of-00007.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:01e6cc7e64b3fca4acef8e01bed4f1208e6179f7cf98d9bd75091192f9662675
|
||||
size 4999813112
|
||||
3
model-00004-of-00007.safetensors
Normal file
3
model-00004-of-00007.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:8eafa6e692a8426e1ab23d10c048dffd209d39034b3a82756bfd81d60a5a4769
|
||||
size 4999813128
|
||||
3
model-00005-of-00007.safetensors
Normal file
3
model-00005-of-00007.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:d56f1c4cc28804ebd05c1db7027f07feaeb13e2ba71b3ebb6d2d549920f431c0
|
||||
size 4832007496
|
||||
3
model-00006-of-00007.safetensors
Normal file
3
model-00006-of-00007.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:60f97bacbfc665b2bd198c024842cb28a55570405e26781a8daa768431230dac
|
||||
size 4999813120
|
||||
3
model-00007-of-00007.safetensors
Normal file
3
model-00007-of-00007.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:b66cfa7710adb2a2e87c7a93cb605850373be16cf37cb223927aaf6f50fea9b7
|
||||
size 2571158184
|
||||
298
model.safetensors.index.json
Normal file
298
model.safetensors.index.json
Normal file
@@ -0,0 +1,298 @@
|
||||
{
|
||||
"metadata": {
|
||||
"total_size": 32121044992
|
||||
},
|
||||
"weight_map": {
|
||||
"lm_head.weight": "model-00007-of-00007.safetensors",
|
||||
"model.embed_tokens.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.input_layernorm.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.mlp.down_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.mlp.gate_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.mlp.up_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.post_attention_layernorm.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.self_attn.k_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.self_attn.o_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.self_attn.q_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.self_attn.v_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.input_layernorm.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.mlp.down_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.mlp.gate_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.mlp.up_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.post_attention_layernorm.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.self_attn.k_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.self_attn.o_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.self_attn.q_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.self_attn.v_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.10.input_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.10.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.10.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.10.mlp.up_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.10.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.10.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.10.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.10.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.10.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.input_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.mlp.up_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.input_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.mlp.up_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.input_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.mlp.up_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.14.input_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.14.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.14.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.14.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.14.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.14.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.14.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.14.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.14.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.15.input_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.15.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.15.mlp.gate_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.15.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.15.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.15.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.15.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.15.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.15.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.input_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.mlp.gate_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.input_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.mlp.gate_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.input_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.mlp.gate_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.input_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.mlp.gate_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.2.input_layernorm.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.2.mlp.down_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.2.mlp.gate_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.2.mlp.up_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.2.post_attention_layernorm.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.2.self_attn.k_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.2.self_attn.o_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.2.self_attn.q_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.2.self_attn.v_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.20.input_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.20.mlp.down_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.20.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.20.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.20.post_attention_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.20.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.20.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.20.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.20.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.21.input_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.21.mlp.down_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.21.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.21.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.21.post_attention_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.21.self_attn.k_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.21.self_attn.o_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.21.self_attn.q_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.21.self_attn.v_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.input_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.mlp.down_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.post_attention_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.self_attn.k_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.self_attn.o_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.self_attn.q_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.self_attn.v_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.input_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.mlp.down_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.post_attention_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.self_attn.k_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.self_attn.o_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.self_attn.q_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.self_attn.v_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.input_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.mlp.down_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.post_attention_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.self_attn.k_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.self_attn.o_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.self_attn.q_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.self_attn.v_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.25.input_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.25.mlp.down_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.25.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.25.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.25.post_attention_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.25.self_attn.k_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.25.self_attn.o_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.25.self_attn.q_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.25.self_attn.v_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.26.input_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.26.mlp.down_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.26.mlp.gate_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.26.mlp.up_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.26.post_attention_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.26.self_attn.k_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.26.self_attn.o_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.26.self_attn.q_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.26.self_attn.v_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.input_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.mlp.down_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.mlp.gate_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.mlp.up_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.post_attention_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.self_attn.k_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.self_attn.o_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.self_attn.q_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.self_attn.v_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.input_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.mlp.down_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.mlp.gate_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.mlp.up_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.post_attention_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.self_attn.k_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.self_attn.o_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.self_attn.q_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.self_attn.v_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.input_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.mlp.down_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.mlp.gate_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.mlp.up_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.post_attention_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.self_attn.k_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.self_attn.o_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.self_attn.q_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.self_attn.v_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.3.input_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.3.mlp.down_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.3.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.3.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.3.post_attention_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.3.self_attn.q_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.3.self_attn.v_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.30.input_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.30.mlp.down_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.30.mlp.gate_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.30.mlp.up_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.30.post_attention_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.30.self_attn.k_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.30.self_attn.o_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.30.self_attn.q_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.30.self_attn.v_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.31.input_layernorm.weight": "model-00007-of-00007.safetensors",
|
||||
"model.layers.31.mlp.down_proj.weight": "model-00007-of-00007.safetensors",
|
||||
"model.layers.31.mlp.gate_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.31.mlp.up_proj.weight": "model-00007-of-00007.safetensors",
|
||||
"model.layers.31.post_attention_layernorm.weight": "model-00007-of-00007.safetensors",
|
||||
"model.layers.31.self_attn.k_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.31.self_attn.o_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.31.self_attn.q_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.31.self_attn.v_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.4.input_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.4.mlp.down_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.4.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.4.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.4.post_attention_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.4.self_attn.k_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.4.self_attn.o_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.4.self_attn.q_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.4.self_attn.v_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.input_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.mlp.down_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.post_attention_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.self_attn.k_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.self_attn.o_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.self_attn.q_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.self_attn.v_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.input_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.mlp.down_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.post_attention_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.self_attn.k_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.self_attn.o_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.self_attn.q_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.self_attn.v_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.input_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.mlp.down_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.post_attention_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.self_attn.k_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.self_attn.o_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.self_attn.q_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.self_attn.v_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.8.input_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.8.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.8.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.8.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.8.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.8.self_attn.k_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.8.self_attn.o_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.8.self_attn.q_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.8.self_attn.v_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.9.input_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.9.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.9.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.9.mlp.up_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.9.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.9.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.9.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.9.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.9.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.norm.weight": "model-00007-of-00007.safetensors"
|
||||
}
|
||||
}
|
||||
23
special_tokens_map.json
Normal file
23
special_tokens_map.json
Normal file
@@ -0,0 +1,23 @@
|
||||
{
|
||||
"bos_token": {
|
||||
"content": "<|begin_of_text|>",
|
||||
"lstrip": false,
|
||||
"normalized": false,
|
||||
"rstrip": false,
|
||||
"single_word": false
|
||||
},
|
||||
"eos_token": {
|
||||
"content": "<|end_of_text|>",
|
||||
"lstrip": false,
|
||||
"normalized": false,
|
||||
"rstrip": false,
|
||||
"single_word": false
|
||||
},
|
||||
"pad_token": {
|
||||
"content": "<|end_of_text|>",
|
||||
"lstrip": false,
|
||||
"normalized": false,
|
||||
"rstrip": false,
|
||||
"single_word": false
|
||||
}
|
||||
}
|
||||
3
tokenizer.json
Normal file
3
tokenizer.json
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:3c5cf44023714fb39b05e71e425f8d7b92805ff73f7988b083b8c87f0bf87393
|
||||
size 17209961
|
||||
2064
tokenizer_config.json
Normal file
2064
tokenizer_config.json
Normal file
File diff suppressed because it is too large
Load Diff
9
train_results.json
Normal file
9
train_results.json
Normal file
@@ -0,0 +1,9 @@
|
||||
{
|
||||
"epoch": 0.9989528795811519,
|
||||
"total_flos": 0.0,
|
||||
"train_loss": 19137.84001572327,
|
||||
"train_runtime": 6782.201,
|
||||
"train_samples": 61135,
|
||||
"train_samples_per_second": 9.014,
|
||||
"train_steps_per_second": 0.07
|
||||
}
|
||||
795
trainer_state.json
Normal file
795
trainer_state.json
Normal file
@@ -0,0 +1,795 @@
|
||||
{
|
||||
"best_global_step": null,
|
||||
"best_metric": null,
|
||||
"best_model_checkpoint": null,
|
||||
"epoch": 0.9989528795811519,
|
||||
"eval_steps": 200,
|
||||
"global_step": 477,
|
||||
"is_hyper_param_search": false,
|
||||
"is_local_process_zero": true,
|
||||
"is_world_process_zero": true,
|
||||
"log_history": [
|
||||
{
|
||||
"epoch": 0.0020942408376963353,
|
||||
"grad_norm": 6690.55859375,
|
||||
"learning_rate": 0.0,
|
||||
"logits/chosen": -0.5995081663131714,
|
||||
"logits/rejected": -0.6144353747367859,
|
||||
"logps/chosen": -1.1853606700897217,
|
||||
"logps/rejected": -1.4816904067993164,
|
||||
"loss": 20000.3594,
|
||||
"rewards/accuracies": 0.59375,
|
||||
"rewards/chosen": 1.3230741387815215e-05,
|
||||
"rewards/margins": 7.880535122239962e-06,
|
||||
"rewards/rejected": 5.350205356080551e-06,
|
||||
"step": 1
|
||||
},
|
||||
{
|
||||
"epoch": 0.020942408376963352,
|
||||
"grad_norm": 5719.5029296875,
|
||||
"learning_rate": 9.375e-08,
|
||||
"logits/chosen": -0.6324554681777954,
|
||||
"logits/rejected": -0.6374175548553467,
|
||||
"logps/chosen": -1.1415884494781494,
|
||||
"logps/rejected": -1.3229657411575317,
|
||||
"loss": 20000.1285,
|
||||
"rewards/accuracies": 0.5243055820465088,
|
||||
"rewards/chosen": -1.3922724519943586e-06,
|
||||
"rewards/margins": -3.091482767558773e-07,
|
||||
"rewards/rejected": -1.0831237204911304e-06,
|
||||
"step": 10
|
||||
},
|
||||
{
|
||||
"epoch": 0.041884816753926704,
|
||||
"grad_norm": 8928.5966796875,
|
||||
"learning_rate": 1.9791666666666664e-07,
|
||||
"logits/chosen": -0.597779393196106,
|
||||
"logits/rejected": -0.6287046670913696,
|
||||
"logps/chosen": -1.1651686429977417,
|
||||
"logps/rejected": -1.311543583869934,
|
||||
"loss": 19999.3844,
|
||||
"rewards/accuracies": 0.5249999761581421,
|
||||
"rewards/chosen": 1.0694350748963188e-05,
|
||||
"rewards/margins": 6.846777978353202e-06,
|
||||
"rewards/rejected": 3.847571861115284e-06,
|
||||
"step": 20
|
||||
},
|
||||
{
|
||||
"epoch": 0.06282722513089005,
|
||||
"grad_norm": 6757.13037109375,
|
||||
"learning_rate": 3.020833333333333e-07,
|
||||
"logits/chosen": -0.6134611368179321,
|
||||
"logits/rejected": -0.605442225933075,
|
||||
"logps/chosen": -1.1174745559692383,
|
||||
"logps/rejected": -1.2003570795059204,
|
||||
"loss": 19997.9344,
|
||||
"rewards/accuracies": 0.5406249761581421,
|
||||
"rewards/chosen": 6.686393317067996e-05,
|
||||
"rewards/margins": 3.225356340408325e-05,
|
||||
"rewards/rejected": 3.461036249063909e-05,
|
||||
"step": 30
|
||||
},
|
||||
{
|
||||
"epoch": 0.08376963350785341,
|
||||
"grad_norm": 8304.0927734375,
|
||||
"learning_rate": 4.0625e-07,
|
||||
"logits/chosen": -0.6177842617034912,
|
||||
"logits/rejected": -0.6428981423377991,
|
||||
"logps/chosen": -1.1158215999603271,
|
||||
"logps/rejected": -1.2773981094360352,
|
||||
"loss": 19990.6594,
|
||||
"rewards/accuracies": 0.612500011920929,
|
||||
"rewards/chosen": 0.0002113355149049312,
|
||||
"rewards/margins": 8.95043122000061e-05,
|
||||
"rewards/rejected": 0.0001218312099808827,
|
||||
"step": 40
|
||||
},
|
||||
{
|
||||
"epoch": 0.10471204188481675,
|
||||
"grad_norm": 6644.90380859375,
|
||||
"learning_rate": 4.999932966293553e-07,
|
||||
"logits/chosen": -0.6271435022354126,
|
||||
"logits/rejected": -0.6673277020454407,
|
||||
"logps/chosen": -1.047498345375061,
|
||||
"logps/rejected": -1.2514413595199585,
|
||||
"loss": 19978.0703,
|
||||
"rewards/accuracies": 0.621874988079071,
|
||||
"rewards/chosen": 0.0006389970076270401,
|
||||
"rewards/margins": 0.0003597235190682113,
|
||||
"rewards/rejected": 0.00027927348855882883,
|
||||
"step": 50
|
||||
},
|
||||
{
|
||||
"epoch": 0.1256544502617801,
|
||||
"grad_norm": 9410.4248046875,
|
||||
"learning_rate": 4.991893270335525e-07,
|
||||
"logits/chosen": -0.6456407308578491,
|
||||
"logits/rejected": -0.6601846814155579,
|
||||
"logps/chosen": -1.1118009090423584,
|
||||
"logps/rejected": -1.2955642938613892,
|
||||
"loss": 19958.8578,
|
||||
"rewards/accuracies": 0.6625000238418579,
|
||||
"rewards/chosen": 0.0007649646140635014,
|
||||
"rewards/margins": 0.00041284109465777874,
|
||||
"rewards/rejected": 0.0003521234611980617,
|
||||
"step": 60
|
||||
},
|
||||
{
|
||||
"epoch": 0.14659685863874344,
|
||||
"grad_norm": 9895.3310546875,
|
||||
"learning_rate": 4.970496218214204e-07,
|
||||
"logits/chosen": -0.710538923740387,
|
||||
"logits/rejected": -0.7164761424064636,
|
||||
"logps/chosen": -1.1343839168548584,
|
||||
"logps/rejected": -1.3560715913772583,
|
||||
"loss": 19939.225,
|
||||
"rewards/accuracies": 0.6499999761581421,
|
||||
"rewards/chosen": 0.00039382357499562204,
|
||||
"rewards/margins": 0.0006254783947952092,
|
||||
"rewards/rejected": -0.0002316548052476719,
|
||||
"step": 70
|
||||
},
|
||||
{
|
||||
"epoch": 0.16753926701570682,
|
||||
"grad_norm": 16232.4453125,
|
||||
"learning_rate": 4.935856505068998e-07,
|
||||
"logits/chosen": -0.6893922090530396,
|
||||
"logits/rejected": -0.6864418387413025,
|
||||
"logps/chosen": -1.1864311695098877,
|
||||
"logps/rejected": -1.511588215827942,
|
||||
"loss": 19887.7375,
|
||||
"rewards/accuracies": 0.675000011920929,
|
||||
"rewards/chosen": -0.0005077069508843124,
|
||||
"rewards/margins": 0.0014377882471308112,
|
||||
"rewards/rejected": -0.0019454952562227845,
|
||||
"step": 80
|
||||
},
|
||||
{
|
||||
"epoch": 0.18848167539267016,
|
||||
"grad_norm": 15383.0810546875,
|
||||
"learning_rate": 4.8881598109976e-07,
|
||||
"logits/chosen": -0.6936720609664917,
|
||||
"logits/rejected": -0.6995416879653931,
|
||||
"logps/chosen": -1.2832069396972656,
|
||||
"logps/rejected": -1.6835159063339233,
|
||||
"loss": 19835.5531,
|
||||
"rewards/accuracies": 0.6625000238418579,
|
||||
"rewards/chosen": -0.0014368193224072456,
|
||||
"rewards/margins": 0.0021624970249831676,
|
||||
"rewards/rejected": -0.003599316580221057,
|
||||
"step": 90
|
||||
},
|
||||
{
|
||||
"epoch": 0.2094240837696335,
|
||||
"grad_norm": 51433.94921875,
|
||||
"learning_rate": 4.827661805750437e-07,
|
||||
"logits/chosen": -0.6618175506591797,
|
||||
"logits/rejected": -0.6787184476852417,
|
||||
"logps/chosen": -1.3208234310150146,
|
||||
"logps/rejected": -1.6875203847885132,
|
||||
"loss": 19789.7328,
|
||||
"rewards/accuracies": 0.668749988079071,
|
||||
"rewards/chosen": -0.001908238627947867,
|
||||
"rewards/margins": 0.0021552981343120337,
|
||||
"rewards/rejected": -0.004063536878675222,
|
||||
"step": 100
|
||||
},
|
||||
{
|
||||
"epoch": 0.23036649214659685,
|
||||
"grad_norm": 39666.109375,
|
||||
"learning_rate": 4.75468677825789e-07,
|
||||
"logits/chosen": -0.6607747077941895,
|
||||
"logits/rejected": -0.6833058595657349,
|
||||
"logps/chosen": -1.3800714015960693,
|
||||
"logps/rejected": -2.1090340614318848,
|
||||
"loss": 19729.8531,
|
||||
"rewards/accuracies": 0.690625011920929,
|
||||
"rewards/chosen": -0.002765479264780879,
|
||||
"rewards/margins": 0.004542567301541567,
|
||||
"rewards/rejected": -0.007308047264814377,
|
||||
"step": 110
|
||||
},
|
||||
{
|
||||
"epoch": 0.2513089005235602,
|
||||
"grad_norm": 26585.208984375,
|
||||
"learning_rate": 4.669625898336438e-07,
|
||||
"logits/chosen": -0.6459494829177856,
|
||||
"logits/rejected": -0.6470843553543091,
|
||||
"logps/chosen": -1.5455596446990967,
|
||||
"logps/rejected": -1.9538530111312866,
|
||||
"loss": 19736.5594,
|
||||
"rewards/accuracies": 0.6312500238418579,
|
||||
"rewards/chosen": -0.004025847185403109,
|
||||
"rewards/margins": 0.0025805418845266104,
|
||||
"rewards/rejected": -0.006606388837099075,
|
||||
"step": 120
|
||||
},
|
||||
{
|
||||
"epoch": 0.27225130890052357,
|
||||
"grad_norm": 39655.12890625,
|
||||
"learning_rate": 4.5729351198915705e-07,
|
||||
"logits/chosen": -0.6332282423973083,
|
||||
"logits/rejected": -0.6239765286445618,
|
||||
"logps/chosen": -1.8061062097549438,
|
||||
"logps/rejected": -2.3397936820983887,
|
||||
"loss": 19656.4844,
|
||||
"rewards/accuracies": 0.6187499761581421,
|
||||
"rewards/chosen": -0.006146098021417856,
|
||||
"rewards/margins": 0.003909106366336346,
|
||||
"rewards/rejected": -0.010055203922092915,
|
||||
"step": 130
|
||||
},
|
||||
{
|
||||
"epoch": 0.2931937172774869,
|
||||
"grad_norm": 63126.5234375,
|
||||
"learning_rate": 4.4651327368569684e-07,
|
||||
"logits/chosen": -0.6292804479598999,
|
||||
"logits/rejected": -0.6358648538589478,
|
||||
"logps/chosen": -2.2388834953308105,
|
||||
"logps/rejected": -2.7073793411254883,
|
||||
"loss": 19726.4313,
|
||||
"rewards/accuracies": 0.578125,
|
||||
"rewards/chosen": -0.01039391104131937,
|
||||
"rewards/margins": 0.003122520400211215,
|
||||
"rewards/rejected": -0.013516431674361229,
|
||||
"step": 140
|
||||
},
|
||||
{
|
||||
"epoch": 0.31413612565445026,
|
||||
"grad_norm": 29782.576171875,
|
||||
"learning_rate": 4.346796604970912e-07,
|
||||
"logits/chosen": -0.6361591815948486,
|
||||
"logits/rejected": -0.635583758354187,
|
||||
"logps/chosen": -2.1060826778411865,
|
||||
"logps/rejected": -3.041602373123169,
|
||||
"loss": 19551.1266,
|
||||
"rewards/accuracies": 0.625,
|
||||
"rewards/chosen": -0.009226142428815365,
|
||||
"rewards/margins": 0.007642732001841068,
|
||||
"rewards/rejected": -0.016868876293301582,
|
||||
"step": 150
|
||||
},
|
||||
{
|
||||
"epoch": 0.33507853403141363,
|
||||
"grad_norm": 37246.7578125,
|
||||
"learning_rate": 4.218561044282098e-07,
|
||||
"logits/chosen": -0.6211342811584473,
|
||||
"logits/rejected": -0.6241915822029114,
|
||||
"logps/chosen": -2.1096720695495605,
|
||||
"logps/rejected": -2.964322566986084,
|
||||
"loss": 19569.3875,
|
||||
"rewards/accuracies": 0.6468750238418579,
|
||||
"rewards/chosen": -0.009453673847019672,
|
||||
"rewards/margins": 0.0064040967263281345,
|
||||
"rewards/rejected": -0.015857771039009094,
|
||||
"step": 160
|
||||
},
|
||||
{
|
||||
"epoch": 0.35602094240837695,
|
||||
"grad_norm": 46280.3671875,
|
||||
"learning_rate": 4.081113438988443e-07,
|
||||
"logits/chosen": -0.6069876551628113,
|
||||
"logits/rejected": -0.6020215749740601,
|
||||
"logps/chosen": -2.012474536895752,
|
||||
"logps/rejected": -3.0016605854034424,
|
||||
"loss": 19453.225,
|
||||
"rewards/accuracies": 0.703125,
|
||||
"rewards/chosen": -0.008933757431805134,
|
||||
"rewards/margins": 0.008303998038172722,
|
||||
"rewards/rejected": -0.01723775453865528,
|
||||
"step": 170
|
||||
},
|
||||
{
|
||||
"epoch": 0.3769633507853403,
|
||||
"grad_norm": 47927.20703125,
|
||||
"learning_rate": 3.935190552834828e-07,
|
||||
"logits/chosen": -0.5999588966369629,
|
||||
"logits/rejected": -0.6140414476394653,
|
||||
"logps/chosen": -2.4049391746520996,
|
||||
"logps/rejected": -3.287382125854492,
|
||||
"loss": 19432.0125,
|
||||
"rewards/accuracies": 0.6156250238418579,
|
||||
"rewards/chosen": -0.012067523784935474,
|
||||
"rewards/margins": 0.007707494311034679,
|
||||
"rewards/rejected": -0.019775018095970154,
|
||||
"step": 180
|
||||
},
|
||||
{
|
||||
"epoch": 0.39790575916230364,
|
||||
"grad_norm": 188406.578125,
|
||||
"learning_rate": 3.781574579820464e-07,
|
||||
"logits/chosen": -0.6130908727645874,
|
||||
"logits/rejected": -0.6305940747261047,
|
||||
"logps/chosen": -2.832430362701416,
|
||||
"logps/rejected": -3.6175262928009033,
|
||||
"loss": 19449.6109,
|
||||
"rewards/accuracies": 0.609375,
|
||||
"rewards/chosen": -0.017013628035783768,
|
||||
"rewards/margins": 0.0060007283464074135,
|
||||
"rewards/rejected": -0.023014355450868607,
|
||||
"step": 190
|
||||
},
|
||||
{
|
||||
"epoch": 0.418848167539267,
|
||||
"grad_norm": 47073.16015625,
|
||||
"learning_rate": 3.621088951385353e-07,
|
||||
"logits/chosen": -0.6221901178359985,
|
||||
"logits/rejected": -0.6376347541809082,
|
||||
"logps/chosen": -2.5921828746795654,
|
||||
"logps/rejected": -3.7902023792266846,
|
||||
"loss": 19275.1328,
|
||||
"rewards/accuracies": 0.609375,
|
||||
"rewards/chosen": -0.01452858280390501,
|
||||
"rewards/margins": 0.01014014147222042,
|
||||
"rewards/rejected": -0.024668725207448006,
|
||||
"step": 200
|
||||
},
|
||||
{
|
||||
"epoch": 0.418848167539267,
|
||||
"eval_logits/chosen": -0.6352167725563049,
|
||||
"eval_logits/rejected": -0.6299898624420166,
|
||||
"eval_logps/chosen": -2.4551029205322266,
|
||||
"eval_logps/rejected": -3.4695351123809814,
|
||||
"eval_loss": 2417.49609375,
|
||||
"eval_rewards/accuracies": 0.656000018119812,
|
||||
"eval_rewards/chosen": -0.013385402970016003,
|
||||
"eval_rewards/margins": 0.00833566952496767,
|
||||
"eval_rewards/rejected": -0.021721070632338524,
|
||||
"eval_runtime": 78.7567,
|
||||
"eval_samples_per_second": 25.395,
|
||||
"eval_steps_per_second": 1.587,
|
||||
"step": 200
|
||||
},
|
||||
{
|
||||
"epoch": 0.4397905759162304,
|
||||
"grad_norm": 27402.61328125,
|
||||
"learning_rate": 3.454593922550693e-07,
|
||||
"logits/chosen": -0.6011684536933899,
|
||||
"logits/rejected": -0.6206346750259399,
|
||||
"logps/chosen": -2.257577419281006,
|
||||
"logps/rejected": -3.6189427375793457,
|
||||
"loss": 19377.5328,
|
||||
"rewards/accuracies": 0.643750011920929,
|
||||
"rewards/chosen": -0.01108148880302906,
|
||||
"rewards/margins": 0.011725915595889091,
|
||||
"rewards/rejected": -0.022807404398918152,
|
||||
"step": 210
|
||||
},
|
||||
{
|
||||
"epoch": 0.4607329842931937,
|
||||
"grad_norm": 55464.56640625,
|
||||
"learning_rate": 3.2829819606729477e-07,
|
||||
"logits/chosen": -0.6579941511154175,
|
||||
"logits/rejected": -0.6557571291923523,
|
||||
"logps/chosen": -2.2686405181884766,
|
||||
"logps/rejected": -3.539849042892456,
|
||||
"loss": 19303.3125,
|
||||
"rewards/accuracies": 0.668749988079071,
|
||||
"rewards/chosen": -0.0108075812458992,
|
||||
"rewards/margins": 0.011197235435247421,
|
||||
"rewards/rejected": -0.02200481668114662,
|
||||
"step": 220
|
||||
},
|
||||
{
|
||||
"epoch": 0.4816753926701571,
|
||||
"grad_norm": 66891.46875,
|
||||
"learning_rate": 3.1071729615293424e-07,
|
||||
"logits/chosen": -0.6366732716560364,
|
||||
"logits/rejected": -0.6232300996780396,
|
||||
"logps/chosen": -2.8289544582366943,
|
||||
"logps/rejected": -3.995779037475586,
|
||||
"loss": 19231.3969,
|
||||
"rewards/accuracies": 0.659375011920929,
|
||||
"rewards/chosen": -0.016343776136636734,
|
||||
"rewards/margins": 0.01012236438691616,
|
||||
"rewards/rejected": -0.026466142386198044,
|
||||
"step": 230
|
||||
},
|
||||
{
|
||||
"epoch": 0.5026178010471204,
|
||||
"grad_norm": 69423.734375,
|
||||
"learning_rate": 2.9281093183781403e-07,
|
||||
"logits/chosen": -0.6015263795852661,
|
||||
"logits/rejected": -0.6166636943817139,
|
||||
"logps/chosen": -3.3009753227233887,
|
||||
"logps/rejected": -4.603926658630371,
|
||||
"loss": 19201.2047,
|
||||
"rewards/accuracies": 0.659375011920929,
|
||||
"rewards/chosen": -0.021539034321904182,
|
||||
"rewards/margins": 0.011443909257650375,
|
||||
"rewards/rejected": -0.03298294544219971,
|
||||
"step": 240
|
||||
},
|
||||
{
|
||||
"epoch": 0.5235602094240838,
|
||||
"grad_norm": 64731.2421875,
|
||||
"learning_rate": 2.7467508704251135e-07,
|
||||
"logits/chosen": -0.6268733739852905,
|
||||
"logits/rejected": -0.6166061162948608,
|
||||
"logps/chosen": -2.9978890419006348,
|
||||
"logps/rejected": -4.9570441246032715,
|
||||
"loss": 18975.8219,
|
||||
"rewards/accuracies": 0.6968749761581421,
|
||||
"rewards/chosen": -0.018680796027183533,
|
||||
"rewards/margins": 0.01662489026784897,
|
||||
"rewards/rejected": -0.0353056825697422,
|
||||
"step": 250
|
||||
},
|
||||
{
|
||||
"epoch": 0.5445026178010471,
|
||||
"grad_norm": 64183.33203125,
|
||||
"learning_rate": 2.5640697577740815e-07,
|
||||
"logits/chosen": -0.6079710721969604,
|
||||
"logits/rejected": -0.5957759618759155,
|
||||
"logps/chosen": -3.0889039039611816,
|
||||
"logps/rejected": -4.906301021575928,
|
||||
"loss": 18967.3656,
|
||||
"rewards/accuracies": 0.6625000238418579,
|
||||
"rewards/chosen": -0.019036870449781418,
|
||||
"rewards/margins": 0.016856301575899124,
|
||||
"rewards/rejected": -0.03589317575097084,
|
||||
"step": 260
|
||||
},
|
||||
{
|
||||
"epoch": 0.5654450261780105,
|
||||
"grad_norm": 103451.53125,
|
||||
"learning_rate": 2.381045210440644e-07,
|
||||
"logits/chosen": -0.5650458335876465,
|
||||
"logits/rejected": -0.5668447613716125,
|
||||
"logps/chosen": -3.9116883277893066,
|
||||
"logps/rejected": -5.719117164611816,
|
||||
"loss": 18905.8031,
|
||||
"rewards/accuracies": 0.625,
|
||||
"rewards/chosen": -0.027667338028550148,
|
||||
"rewards/margins": 0.016292227432131767,
|
||||
"rewards/rejected": -0.043959565460681915,
|
||||
"step": 270
|
||||
},
|
||||
{
|
||||
"epoch": 0.5863874345549738,
|
||||
"grad_norm": 74060.140625,
|
||||
"learning_rate": 2.1986582993616925e-07,
|
||||
"logits/chosen": -0.562662661075592,
|
||||
"logits/rejected": -0.5547593235969543,
|
||||
"logps/chosen": -3.6009249687194824,
|
||||
"logps/rejected": -6.734810829162598,
|
||||
"loss": 18614.3531,
|
||||
"rewards/accuracies": 0.7124999761581421,
|
||||
"rewards/chosen": -0.02406414784491062,
|
||||
"rewards/margins": 0.029486840590834618,
|
||||
"rewards/rejected": -0.05355098843574524,
|
||||
"step": 280
|
||||
},
|
||||
{
|
||||
"epoch": 0.6073298429319371,
|
||||
"grad_norm": 197744.75,
|
||||
"learning_rate": 2.0178866775369774e-07,
|
||||
"logits/chosen": -0.5442657470703125,
|
||||
"logits/rejected": -0.5202258825302124,
|
||||
"logps/chosen": -4.191218852996826,
|
||||
"logps/rejected": -5.84238862991333,
|
||||
"loss": 18823.9969,
|
||||
"rewards/accuracies": 0.596875011920929,
|
||||
"rewards/chosen": -0.0303532425314188,
|
||||
"rewards/margins": 0.014872364699840546,
|
||||
"rewards/rejected": -0.045225612819194794,
|
||||
"step": 290
|
||||
},
|
||||
{
|
||||
"epoch": 0.6282722513089005,
|
||||
"grad_norm": 82288.1015625,
|
||||
"learning_rate": 1.839699339491937e-07,
|
||||
"logits/chosen": -0.5231214165687561,
|
||||
"logits/rejected": -0.5048767328262329,
|
||||
"logps/chosen": -4.525745868682861,
|
||||
"logps/rejected": -5.8891825675964355,
|
||||
"loss": 18956.8344,
|
||||
"rewards/accuracies": 0.6343749761581421,
|
||||
"rewards/chosen": -0.03319484740495682,
|
||||
"rewards/margins": 0.013094061985611916,
|
||||
"rewards/rejected": -0.046288907527923584,
|
||||
"step": 300
|
||||
},
|
||||
{
|
||||
"epoch": 0.6492146596858639,
|
||||
"grad_norm": 83390.5,
|
||||
"learning_rate": 1.6650514271527465e-07,
|
||||
"logits/chosen": -0.47909316420555115,
|
||||
"logits/rejected": -0.4821901321411133,
|
||||
"logps/chosen": -3.890228271484375,
|
||||
"logps/rejected": -5.4383721351623535,
|
||||
"loss": 18785.9547,
|
||||
"rewards/accuracies": 0.671875,
|
||||
"rewards/chosen": -0.027516454458236694,
|
||||
"rewards/margins": 0.014197492972016335,
|
||||
"rewards/rejected": -0.04171394929289818,
|
||||
"step": 310
|
||||
},
|
||||
{
|
||||
"epoch": 0.6701570680628273,
|
||||
"grad_norm": 119347.234375,
|
||||
"learning_rate": 1.4948791099758052e-07,
|
||||
"logits/chosen": -0.47245222330093384,
|
||||
"logits/rejected": -0.48146170377731323,
|
||||
"logps/chosen": -4.161627769470215,
|
||||
"logps/rejected": -6.324838161468506,
|
||||
"loss": 18573.5031,
|
||||
"rewards/accuracies": 0.6781250238418579,
|
||||
"rewards/chosen": -0.030029457062482834,
|
||||
"rewards/margins": 0.020363491028547287,
|
||||
"rewards/rejected": -0.05039294809103012,
|
||||
"step": 320
|
||||
},
|
||||
{
|
||||
"epoch": 0.6910994764397905,
|
||||
"grad_norm": 108725.671875,
|
||||
"learning_rate": 1.3300945667758012e-07,
|
||||
"logits/chosen": -0.4325336515903473,
|
||||
"logits/rejected": -0.43659958243370056,
|
||||
"logps/chosen": -4.2359724044799805,
|
||||
"logps/rejected": -7.189882755279541,
|
||||
"loss": 18710.9344,
|
||||
"rewards/accuracies": 0.6781250238418579,
|
||||
"rewards/chosen": -0.03000812791287899,
|
||||
"rewards/margins": 0.028335105627775192,
|
||||
"rewards/rejected": -0.05834323167800903,
|
||||
"step": 330
|
||||
},
|
||||
{
|
||||
"epoch": 0.7120418848167539,
|
||||
"grad_norm": 250133.859375,
|
||||
"learning_rate": 1.1715810961514072e-07,
|
||||
"logits/chosen": -0.40766844153404236,
|
||||
"logits/rejected": -0.40503817796707153,
|
||||
"logps/chosen": -5.0073561668396,
|
||||
"logps/rejected": -7.256580352783203,
|
||||
"loss": 18789.4781,
|
||||
"rewards/accuracies": 0.6875,
|
||||
"rewards/chosen": -0.037794455885887146,
|
||||
"rewards/margins": 0.021081697195768356,
|
||||
"rewards/rejected": -0.0588761568069458,
|
||||
"step": 340
|
||||
},
|
||||
{
|
||||
"epoch": 0.7329842931937173,
|
||||
"grad_norm": 65695.078125,
|
||||
"learning_rate": 1.0201883817182949e-07,
|
||||
"logits/chosen": -0.402383416891098,
|
||||
"logits/rejected": -0.4145272374153137,
|
||||
"logps/chosen": -4.494850158691406,
|
||||
"logps/rejected": -6.888899803161621,
|
||||
"loss": 18565.8484,
|
||||
"rewards/accuracies": 0.6968749761581421,
|
||||
"rewards/chosen": -0.03312065079808235,
|
||||
"rewards/margins": 0.0219460166990757,
|
||||
"rewards/rejected": -0.05506666377186775,
|
||||
"step": 350
|
||||
},
|
||||
{
|
||||
"epoch": 0.7539267015706806,
|
||||
"grad_norm": 182331.0625,
|
||||
"learning_rate": 8.76727937529367e-08,
|
||||
"logits/chosen": -0.3540952503681183,
|
||||
"logits/rejected": -0.36194995045661926,
|
||||
"logps/chosen": -4.524051189422607,
|
||||
"logps/rejected": -7.192444801330566,
|
||||
"loss": 18681.9938,
|
||||
"rewards/accuracies": 0.628125011920929,
|
||||
"rewards/chosen": -0.03363392874598503,
|
||||
"rewards/margins": 0.024621980264782906,
|
||||
"rewards/rejected": -0.058255910873413086,
|
||||
"step": 360
|
||||
},
|
||||
{
|
||||
"epoch": 0.774869109947644,
|
||||
"grad_norm": 99043.015625,
|
||||
"learning_rate": 7.419687580962222e-08,
|
||||
"logits/chosen": -0.38287925720214844,
|
||||
"logits/rejected": -0.3868221640586853,
|
||||
"logps/chosen": -4.889115810394287,
|
||||
"logps/rejected": -7.343722343444824,
|
||||
"loss": 18350.1594,
|
||||
"rewards/accuracies": 0.6625000238418579,
|
||||
"rewards/chosen": -0.036828476935625076,
|
||||
"rewards/margins": 0.02258094772696495,
|
||||
"rewards/rejected": -0.059409428387880325,
|
||||
"step": 370
|
||||
},
|
||||
{
|
||||
"epoch": 0.7958115183246073,
|
||||
"grad_norm": 118281.1953125,
|
||||
"learning_rate": 6.166331963291519e-08,
|
||||
"logits/chosen": -0.29700514674186707,
|
||||
"logits/rejected": -0.3104197382926941,
|
||||
"logps/chosen": -4.535180568695068,
|
||||
"logps/rejected": -7.199211120605469,
|
||||
"loss": 18479.7844,
|
||||
"rewards/accuracies": 0.699999988079071,
|
||||
"rewards/chosen": -0.033621106296777725,
|
||||
"rewards/margins": 0.02514580450952053,
|
||||
"rewards/rejected": -0.058766912668943405,
|
||||
"step": 380
|
||||
},
|
||||
{
|
||||
"epoch": 0.8167539267015707,
|
||||
"grad_norm": 116531.1015625,
|
||||
"learning_rate": 5.013930914912476e-08,
|
||||
"logits/chosen": -0.28081822395324707,
|
||||
"logits/rejected": -0.29538464546203613,
|
||||
"logps/chosen": -5.330414772033691,
|
||||
"logps/rejected": -7.864521026611328,
|
||||
"loss": 18660.3094,
|
||||
"rewards/accuracies": 0.668749988079071,
|
||||
"rewards/chosen": -0.041854970157146454,
|
||||
"rewards/margins": 0.023673858493566513,
|
||||
"rewards/rejected": -0.06552883237600327,
|
||||
"step": 390
|
||||
},
|
||||
{
|
||||
"epoch": 0.837696335078534,
|
||||
"grad_norm": 115313.1640625,
|
||||
"learning_rate": 3.968661679220467e-08,
|
||||
"logits/chosen": -0.2553872764110565,
|
||||
"logits/rejected": -0.27801090478897095,
|
||||
"logps/chosen": -5.410425662994385,
|
||||
"logps/rejected": -7.3440046310424805,
|
||||
"loss": 18486.2438,
|
||||
"rewards/accuracies": 0.637499988079071,
|
||||
"rewards/chosen": -0.04237430915236473,
|
||||
"rewards/margins": 0.018501609563827515,
|
||||
"rewards/rejected": -0.060875922441482544,
|
||||
"step": 400
|
||||
},
|
||||
{
|
||||
"epoch": 0.837696335078534,
|
||||
"eval_logits/chosen": -0.2619972229003906,
|
||||
"eval_logits/rejected": -0.27771249413490295,
|
||||
"eval_logps/chosen": -5.502514839172363,
|
||||
"eval_logps/rejected": -8.184652328491211,
|
||||
"eval_loss": 2313.8056640625,
|
||||
"eval_rewards/accuracies": 0.6800000071525574,
|
||||
"eval_rewards/chosen": -0.04385951906442642,
|
||||
"eval_rewards/margins": 0.025012729689478874,
|
||||
"eval_rewards/rejected": -0.06887225061655045,
|
||||
"eval_runtime": 78.5743,
|
||||
"eval_samples_per_second": 25.454,
|
||||
"eval_steps_per_second": 1.591,
|
||||
"step": 400
|
||||
},
|
||||
{
|
||||
"epoch": 0.8586387434554974,
|
||||
"grad_norm": 116385.09375,
|
||||
"learning_rate": 3.036127238347164e-08,
|
||||
"logits/chosen": -0.24211814999580383,
|
||||
"logits/rejected": -0.2585967183113098,
|
||||
"logps/chosen": -5.910313606262207,
|
||||
"logps/rejected": -8.293985366821289,
|
||||
"loss": 18503.9844,
|
||||
"rewards/accuracies": 0.6343749761581421,
|
||||
"rewards/chosen": -0.047523993998765945,
|
||||
"rewards/margins": 0.02237233892083168,
|
||||
"rewards/rejected": -0.06989633291959763,
|
||||
"step": 410
|
||||
},
|
||||
{
|
||||
"epoch": 0.8795811518324608,
|
||||
"grad_norm": 134170.890625,
|
||||
"learning_rate": 2.2213262793589482e-08,
|
||||
"logits/chosen": -0.22771398723125458,
|
||||
"logits/rejected": -0.2378026694059372,
|
||||
"logps/chosen": -5.347130298614502,
|
||||
"logps/rejected": -8.04057788848877,
|
||||
"loss": 18328.1453,
|
||||
"rewards/accuracies": 0.671875,
|
||||
"rewards/chosen": -0.042228274047374725,
|
||||
"rewards/margins": 0.025333663448691368,
|
||||
"rewards/rejected": -0.06756193935871124,
|
||||
"step": 420
|
||||
},
|
||||
{
|
||||
"epoch": 0.900523560209424,
|
||||
"grad_norm": 140065.46875,
|
||||
"learning_rate": 1.5286263996730026e-08,
|
||||
"logits/chosen": -0.22656838595867157,
|
||||
"logits/rejected": -0.26688051223754883,
|
||||
"logps/chosen": -5.213972568511963,
|
||||
"logps/rejected": -8.997479438781738,
|
||||
"loss": 18304.7516,
|
||||
"rewards/accuracies": 0.6781250238418579,
|
||||
"rewards/chosen": -0.040835700929164886,
|
||||
"rewards/margins": 0.03499193489551544,
|
||||
"rewards/rejected": -0.07582763582468033,
|
||||
"step": 430
|
||||
},
|
||||
{
|
||||
"epoch": 0.9214659685863874,
|
||||
"grad_norm": 147417.296875,
|
||||
"learning_rate": 9.617406953185136e-09,
|
||||
"logits/chosen": -0.21701665222644806,
|
||||
"logits/rejected": -0.22004380822181702,
|
||||
"logps/chosen": -5.577446937561035,
|
||||
"logps/rejected": -8.317059516906738,
|
||||
"loss": 18362.8578,
|
||||
"rewards/accuracies": 0.643750011920929,
|
||||
"rewards/chosen": -0.04439837858080864,
|
||||
"rewards/margins": 0.025850927457213402,
|
||||
"rewards/rejected": -0.07024930417537689,
|
||||
"step": 440
|
||||
},
|
||||
{
|
||||
"epoch": 0.9424083769633508,
|
||||
"grad_norm": 187662.15625,
|
||||
"learning_rate": 5.2370785753763356e-09,
|
||||
"logits/chosen": -0.19742074608802795,
|
||||
"logits/rejected": -0.21276791393756866,
|
||||
"logps/chosen": -5.669443607330322,
|
||||
"logps/rejected": -8.552831649780273,
|
||||
"loss": 18458.9266,
|
||||
"rewards/accuracies": 0.6812499761581421,
|
||||
"rewards/chosen": -0.0450492687523365,
|
||||
"rewards/margins": 0.027155417948961258,
|
||||
"rewards/rejected": -0.07220469415187836,
|
||||
"step": 450
|
||||
},
|
||||
{
|
||||
"epoch": 0.9633507853403142,
|
||||
"grad_norm": 124422.875,
|
||||
"learning_rate": 2.168758844148272e-09,
|
||||
"logits/chosen": -0.18504954874515533,
|
||||
"logits/rejected": -0.22272753715515137,
|
||||
"logps/chosen": -5.632592678070068,
|
||||
"logps/rejected": -8.428214073181152,
|
||||
"loss": 18551.7969,
|
||||
"rewards/accuracies": 0.6781250238418579,
|
||||
"rewards/chosen": -0.04460041597485542,
|
||||
"rewards/margins": 0.026442039757966995,
|
||||
"rewards/rejected": -0.07104245573282242,
|
||||
"step": 460
|
||||
},
|
||||
{
|
||||
"epoch": 0.9842931937172775,
|
||||
"grad_norm": 157096.296875,
|
||||
"learning_rate": 4.288949484559934e-10,
|
||||
"logits/chosen": -0.14469654858112335,
|
||||
"logits/rejected": -0.15391038358211517,
|
||||
"logps/chosen": -5.636763572692871,
|
||||
"logps/rejected": -8.912847518920898,
|
||||
"loss": 18241.4094,
|
||||
"rewards/accuracies": 0.6812499761581421,
|
||||
"rewards/chosen": -0.04426239803433418,
|
||||
"rewards/margins": 0.031570300459861755,
|
||||
"rewards/rejected": -0.07583270221948624,
|
||||
"step": 470
|
||||
},
|
||||
{
|
||||
"epoch": 0.9989528795811519,
|
||||
"step": 477,
|
||||
"total_flos": 0.0,
|
||||
"train_loss": 19137.84001572327,
|
||||
"train_runtime": 6782.201,
|
||||
"train_samples_per_second": 9.014,
|
||||
"train_steps_per_second": 0.07
|
||||
}
|
||||
],
|
||||
"logging_steps": 10,
|
||||
"max_steps": 477,
|
||||
"num_input_tokens_seen": 0,
|
||||
"num_train_epochs": 1,
|
||||
"save_steps": 200,
|
||||
"stateful_callbacks": {
|
||||
"TrainerControl": {
|
||||
"args": {
|
||||
"should_epoch_stop": false,
|
||||
"should_evaluate": false,
|
||||
"should_log": false,
|
||||
"should_save": true,
|
||||
"should_training_stop": true
|
||||
},
|
||||
"attributes": {}
|
||||
}
|
||||
},
|
||||
"total_flos": 0.0,
|
||||
"train_batch_size": 4,
|
||||
"trial_name": null,
|
||||
"trial_params": null
|
||||
}
|
||||
Reference in New Issue
Block a user