初始化项目,由ModelHub XC社区提供模型
Model: jackf857/llama-3-8b-base-robust-dpo-ultrafeedback-8xh200 Source: Original Platform
This commit is contained in:
36
.gitattributes
vendored
Normal file
36
.gitattributes
vendored
Normal file
@@ -0,0 +1,36 @@
|
||||
*.7z filter=lfs diff=lfs merge=lfs -text
|
||||
*.arrow filter=lfs diff=lfs merge=lfs -text
|
||||
*.bin filter=lfs diff=lfs merge=lfs -text
|
||||
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
||||
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
||||
*.ftz filter=lfs diff=lfs merge=lfs -text
|
||||
*.gz filter=lfs diff=lfs merge=lfs -text
|
||||
*.h5 filter=lfs diff=lfs merge=lfs -text
|
||||
*.joblib filter=lfs diff=lfs merge=lfs -text
|
||||
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
||||
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
||||
*.model filter=lfs diff=lfs merge=lfs -text
|
||||
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
||||
*.npy filter=lfs diff=lfs merge=lfs -text
|
||||
*.npz filter=lfs diff=lfs merge=lfs -text
|
||||
*.onnx filter=lfs diff=lfs merge=lfs -text
|
||||
*.ot filter=lfs diff=lfs merge=lfs -text
|
||||
*.parquet filter=lfs diff=lfs merge=lfs -text
|
||||
*.pb filter=lfs diff=lfs merge=lfs -text
|
||||
*.pickle filter=lfs diff=lfs merge=lfs -text
|
||||
*.pkl filter=lfs diff=lfs merge=lfs -text
|
||||
*.pt filter=lfs diff=lfs merge=lfs -text
|
||||
*.pth filter=lfs diff=lfs merge=lfs -text
|
||||
*.rar filter=lfs diff=lfs merge=lfs -text
|
||||
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
||||
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
||||
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
||||
*.tar filter=lfs diff=lfs merge=lfs -text
|
||||
*.tflite filter=lfs diff=lfs merge=lfs -text
|
||||
*.tgz filter=lfs diff=lfs merge=lfs -text
|
||||
*.wasm filter=lfs diff=lfs merge=lfs -text
|
||||
*.xz filter=lfs diff=lfs merge=lfs -text
|
||||
*.zip filter=lfs diff=lfs merge=lfs -text
|
||||
*.zst filter=lfs diff=lfs merge=lfs -text
|
||||
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
||||
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
||||
76
README.md
Normal file
76
README.md
Normal file
@@ -0,0 +1,76 @@
|
||||
---
|
||||
library_name: transformers
|
||||
base_model: W-61/llama-3-8b-base-sft-ultrachat-8xh200
|
||||
tags:
|
||||
- alignment-handbook
|
||||
- robust-dpo
|
||||
- generated_from_trainer
|
||||
datasets:
|
||||
- HuggingFaceH4/ultrafeedback_binarized
|
||||
model-index:
|
||||
- name: llama-3-8b-base-robust-dpo-ultrafeedback-8xh200
|
||||
results: []
|
||||
---
|
||||
|
||||
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
||||
should probably proofread and complete it, then remove this comment. -->
|
||||
|
||||
# llama-3-8b-base-robust-dpo-ultrafeedback-8xh200
|
||||
|
||||
This model is a fine-tuned version of [W-61/llama-3-8b-base-sft-ultrachat-8xh200](https://huggingface.co/W-61/llama-3-8b-base-sft-ultrachat-8xh200) on the HuggingFaceH4/ultrafeedback_binarized dataset.
|
||||
It achieves the following results on the evaluation set:
|
||||
- Loss: 0.3504
|
||||
- Rewards/chosen: -0.2669
|
||||
- Rewards/rejected: -1.9046
|
||||
- Rewards/accuracies: 0.7460
|
||||
- Rewards/margins: 1.6376
|
||||
- Logps/rejected: -286.3951
|
||||
- Logps/chosen: -300.5437
|
||||
- Logits/rejected: -0.7660
|
||||
- Logits/chosen: -0.7842
|
||||
|
||||
## Model description
|
||||
|
||||
More information needed
|
||||
|
||||
## Intended uses & limitations
|
||||
|
||||
More information needed
|
||||
|
||||
## Training and evaluation data
|
||||
|
||||
More information needed
|
||||
|
||||
## Training procedure
|
||||
|
||||
### Training hyperparameters
|
||||
|
||||
The following hyperparameters were used during training:
|
||||
- learning_rate: 5e-07
|
||||
- train_batch_size: 4
|
||||
- eval_batch_size: 4
|
||||
- seed: 42
|
||||
- distributed_type: multi-GPU
|
||||
- num_devices: 8
|
||||
- gradient_accumulation_steps: 4
|
||||
- total_train_batch_size: 128
|
||||
- total_eval_batch_size: 32
|
||||
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
||||
- lr_scheduler_type: cosine
|
||||
- lr_scheduler_warmup_ratio: 0.1
|
||||
- num_epochs: 1
|
||||
|
||||
### Training results
|
||||
|
||||
| Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
|
||||
|:-------------:|:------:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
|
||||
| 1.5682 | 0.4188 | 200 | 0.4018 | -0.2422 | -1.5310 | 0.7540 | 1.2888 | -282.6593 | -300.2964 | -0.7831 | -0.8001 |
|
||||
| 1.3853 | 0.8377 | 400 | 0.3504 | -0.2669 | -1.9046 | 0.7460 | 1.6376 | -286.3951 | -300.5437 | -0.7660 | -0.7842 |
|
||||
|
||||
|
||||
### Framework versions
|
||||
|
||||
- Transformers 4.51.0
|
||||
- Pytorch 2.3.1+cu121
|
||||
- Datasets 2.21.0
|
||||
- Tokenizers 0.21.4
|
||||
22
all_results.json
Normal file
22
all_results.json
Normal file
@@ -0,0 +1,22 @@
|
||||
{
|
||||
"epoch": 0.9989528795811519,
|
||||
"eval_logits/chosen": -0.790234386920929,
|
||||
"eval_logits/rejected": -0.7727676630020142,
|
||||
"eval_logps/chosen": -301.2696838378906,
|
||||
"eval_logps/rejected": -287.254638671875,
|
||||
"eval_loss": 0.347857266664505,
|
||||
"eval_rewards/accuracies": 0.7580645084381104,
|
||||
"eval_rewards/chosen": -0.33950960636138916,
|
||||
"eval_rewards/margins": 1.651003122329712,
|
||||
"eval_rewards/rejected": -1.990512728691101,
|
||||
"eval_runtime": 44.5446,
|
||||
"eval_samples": 2000,
|
||||
"eval_samples_per_second": 44.899,
|
||||
"eval_steps_per_second": 1.414,
|
||||
"total_flos": 0.0,
|
||||
"train_loss": 1.6700679651096433,
|
||||
"train_runtime": 3395.1223,
|
||||
"train_samples": 61135,
|
||||
"train_samples_per_second": 18.007,
|
||||
"train_steps_per_second": 0.14
|
||||
}
|
||||
29
config.json
Normal file
29
config.json
Normal file
@@ -0,0 +1,29 @@
|
||||
{
|
||||
"architectures": [
|
||||
"LlamaForCausalLM"
|
||||
],
|
||||
"attention_bias": false,
|
||||
"attention_dropout": 0.0,
|
||||
"bos_token_id": 128000,
|
||||
"eos_token_id": 128001,
|
||||
"head_dim": 128,
|
||||
"hidden_act": "silu",
|
||||
"hidden_size": 4096,
|
||||
"initializer_range": 0.02,
|
||||
"intermediate_size": 14336,
|
||||
"max_position_embeddings": 8192,
|
||||
"mlp_bias": false,
|
||||
"model_type": "llama",
|
||||
"num_attention_heads": 32,
|
||||
"num_hidden_layers": 32,
|
||||
"num_key_value_heads": 8,
|
||||
"pretraining_tp": 1,
|
||||
"rms_norm_eps": 1e-05,
|
||||
"rope_scaling": null,
|
||||
"rope_theta": 500000.0,
|
||||
"tie_word_embeddings": false,
|
||||
"torch_dtype": "float32",
|
||||
"transformers_version": "4.51.0",
|
||||
"use_cache": true,
|
||||
"vocab_size": 128256
|
||||
}
|
||||
16
eval_results.json
Normal file
16
eval_results.json
Normal file
@@ -0,0 +1,16 @@
|
||||
{
|
||||
"epoch": 0.9989528795811519,
|
||||
"eval_logits/chosen": -0.790234386920929,
|
||||
"eval_logits/rejected": -0.7727676630020142,
|
||||
"eval_logps/chosen": -301.2696838378906,
|
||||
"eval_logps/rejected": -287.254638671875,
|
||||
"eval_loss": 0.347857266664505,
|
||||
"eval_rewards/accuracies": 0.7580645084381104,
|
||||
"eval_rewards/chosen": -0.33950960636138916,
|
||||
"eval_rewards/margins": 1.651003122329712,
|
||||
"eval_rewards/rejected": -1.990512728691101,
|
||||
"eval_runtime": 44.5446,
|
||||
"eval_samples": 2000,
|
||||
"eval_samples_per_second": 44.899,
|
||||
"eval_steps_per_second": 1.414
|
||||
}
|
||||
9
generation_config.json
Normal file
9
generation_config.json
Normal file
@@ -0,0 +1,9 @@
|
||||
{
|
||||
"bos_token_id": 128000,
|
||||
"do_sample": true,
|
||||
"eos_token_id": 128001,
|
||||
"max_length": 4096,
|
||||
"temperature": 0.6,
|
||||
"top_p": 0.9,
|
||||
"transformers_version": "4.51.0"
|
||||
}
|
||||
3
model-00001-of-00007.safetensors
Normal file
3
model-00001-of-00007.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:c0222633cecdcfc94e79513851714ec69dfb1a8b537c108ebb5fb96577a9f28a
|
||||
size 4886466168
|
||||
3
model-00002-of-00007.safetensors
Normal file
3
model-00002-of-00007.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:a57afd7965bf5d3f35259890367381a001b6771c1286e47be41b00ec6fc4042b
|
||||
size 4832007448
|
||||
3
model-00003-of-00007.safetensors
Normal file
3
model-00003-of-00007.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:28e2cf60fedcb3b2cb82d688a5f21a5cb7df4f5858faddf10410cf5c2173b4f5
|
||||
size 4999813112
|
||||
3
model-00004-of-00007.safetensors
Normal file
3
model-00004-of-00007.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:9f7ae689b17ebebfcf658bd914e1e9eca1b7aaa8180f58706db3b95dfbbb3f79
|
||||
size 4999813128
|
||||
3
model-00005-of-00007.safetensors
Normal file
3
model-00005-of-00007.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:7e6ae58355fcf64feb25a51e5f0e2d288b0419af7d6bcf87c35a4daa0f149f12
|
||||
size 4832007496
|
||||
3
model-00006-of-00007.safetensors
Normal file
3
model-00006-of-00007.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:e78ae755d296c02acab23705ac951219c73defb568f3475ecfbd781023b3e301
|
||||
size 4999813120
|
||||
3
model-00007-of-00007.safetensors
Normal file
3
model-00007-of-00007.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:0c8b46018574add646616317e5369ff8b1a6771e4a0b7803bae40ab1430ea391
|
||||
size 2571158184
|
||||
298
model.safetensors.index.json
Normal file
298
model.safetensors.index.json
Normal file
@@ -0,0 +1,298 @@
|
||||
{
|
||||
"metadata": {
|
||||
"total_size": 32121044992
|
||||
},
|
||||
"weight_map": {
|
||||
"lm_head.weight": "model-00007-of-00007.safetensors",
|
||||
"model.embed_tokens.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.input_layernorm.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.mlp.down_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.mlp.gate_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.mlp.up_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.post_attention_layernorm.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.self_attn.k_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.self_attn.o_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.self_attn.q_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.0.self_attn.v_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.input_layernorm.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.mlp.down_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.mlp.gate_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.mlp.up_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.post_attention_layernorm.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.self_attn.k_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.self_attn.o_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.self_attn.q_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.1.self_attn.v_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.10.input_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.10.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.10.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.10.mlp.up_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.10.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.10.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.10.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.10.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.10.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.input_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.mlp.up_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.11.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.input_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.mlp.up_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.12.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.input_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.mlp.up_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.13.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.14.input_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.14.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.14.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.14.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.14.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.14.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.14.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.14.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.14.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.15.input_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.15.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.15.mlp.gate_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.15.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.15.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.15.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.15.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.15.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.15.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.input_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.mlp.gate_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.16.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.input_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.mlp.gate_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.17.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.input_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.mlp.gate_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.18.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.input_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.mlp.gate_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.19.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.2.input_layernorm.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.2.mlp.down_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.2.mlp.gate_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.2.mlp.up_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.2.post_attention_layernorm.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.2.self_attn.k_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.2.self_attn.o_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.2.self_attn.q_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.2.self_attn.v_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.20.input_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.20.mlp.down_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.20.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.20.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.20.post_attention_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.20.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.20.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.20.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.20.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
|
||||
"model.layers.21.input_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.21.mlp.down_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.21.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.21.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.21.post_attention_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.21.self_attn.k_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.21.self_attn.o_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.21.self_attn.q_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.21.self_attn.v_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.input_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.mlp.down_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.post_attention_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.self_attn.k_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.self_attn.o_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.self_attn.q_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.22.self_attn.v_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.input_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.mlp.down_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.post_attention_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.self_attn.k_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.self_attn.o_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.self_attn.q_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.23.self_attn.v_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.input_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.mlp.down_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.post_attention_layernorm.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.self_attn.k_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.self_attn.o_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.self_attn.q_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.24.self_attn.v_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.25.input_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.25.mlp.down_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.25.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.25.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.25.post_attention_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.25.self_attn.k_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.25.self_attn.o_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.25.self_attn.q_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.25.self_attn.v_proj.weight": "model-00005-of-00007.safetensors",
|
||||
"model.layers.26.input_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.26.mlp.down_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.26.mlp.gate_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.26.mlp.up_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.26.post_attention_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.26.self_attn.k_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.26.self_attn.o_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.26.self_attn.q_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.26.self_attn.v_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.input_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.mlp.down_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.mlp.gate_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.mlp.up_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.post_attention_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.self_attn.k_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.self_attn.o_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.self_attn.q_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.27.self_attn.v_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.input_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.mlp.down_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.mlp.gate_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.mlp.up_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.post_attention_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.self_attn.k_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.self_attn.o_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.self_attn.q_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.28.self_attn.v_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.input_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.mlp.down_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.mlp.gate_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.mlp.up_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.post_attention_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.self_attn.k_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.self_attn.o_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.self_attn.q_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.29.self_attn.v_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.3.input_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.3.mlp.down_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.3.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.3.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.3.post_attention_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.3.self_attn.q_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.3.self_attn.v_proj.weight": "model-00001-of-00007.safetensors",
|
||||
"model.layers.30.input_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.30.mlp.down_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.30.mlp.gate_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.30.mlp.up_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.30.post_attention_layernorm.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.30.self_attn.k_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.30.self_attn.o_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.30.self_attn.q_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.30.self_attn.v_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.31.input_layernorm.weight": "model-00007-of-00007.safetensors",
|
||||
"model.layers.31.mlp.down_proj.weight": "model-00007-of-00007.safetensors",
|
||||
"model.layers.31.mlp.gate_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.31.mlp.up_proj.weight": "model-00007-of-00007.safetensors",
|
||||
"model.layers.31.post_attention_layernorm.weight": "model-00007-of-00007.safetensors",
|
||||
"model.layers.31.self_attn.k_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.31.self_attn.o_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.31.self_attn.q_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.31.self_attn.v_proj.weight": "model-00006-of-00007.safetensors",
|
||||
"model.layers.4.input_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.4.mlp.down_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.4.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.4.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.4.post_attention_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.4.self_attn.k_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.4.self_attn.o_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.4.self_attn.q_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.4.self_attn.v_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.input_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.mlp.down_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.post_attention_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.self_attn.k_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.self_attn.o_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.self_attn.q_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.5.self_attn.v_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.input_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.mlp.down_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.post_attention_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.self_attn.k_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.self_attn.o_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.self_attn.q_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.6.self_attn.v_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.input_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.mlp.down_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.post_attention_layernorm.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.self_attn.k_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.self_attn.o_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.self_attn.q_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.7.self_attn.v_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.8.input_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.8.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.8.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.8.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.8.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.8.self_attn.k_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.8.self_attn.o_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.8.self_attn.q_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.8.self_attn.v_proj.weight": "model-00002-of-00007.safetensors",
|
||||
"model.layers.9.input_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.9.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.9.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.9.mlp.up_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.9.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.9.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.9.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.9.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.layers.9.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
|
||||
"model.norm.weight": "model-00007-of-00007.safetensors"
|
||||
}
|
||||
}
|
||||
23
special_tokens_map.json
Normal file
23
special_tokens_map.json
Normal file
@@ -0,0 +1,23 @@
|
||||
{
|
||||
"bos_token": {
|
||||
"content": "<|begin_of_text|>",
|
||||
"lstrip": false,
|
||||
"normalized": false,
|
||||
"rstrip": false,
|
||||
"single_word": false
|
||||
},
|
||||
"eos_token": {
|
||||
"content": "<|end_of_text|>",
|
||||
"lstrip": false,
|
||||
"normalized": false,
|
||||
"rstrip": false,
|
||||
"single_word": false
|
||||
},
|
||||
"pad_token": {
|
||||
"content": "<|end_of_text|>",
|
||||
"lstrip": false,
|
||||
"normalized": false,
|
||||
"rstrip": false,
|
||||
"single_word": false
|
||||
}
|
||||
}
|
||||
3
tokenizer.json
Normal file
3
tokenizer.json
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:3c5cf44023714fb39b05e71e425f8d7b92805ff73f7988b083b8c87f0bf87393
|
||||
size 17209961
|
||||
2064
tokenizer_config.json
Normal file
2064
tokenizer_config.json
Normal file
File diff suppressed because it is too large
Load Diff
9
train_results.json
Normal file
9
train_results.json
Normal file
@@ -0,0 +1,9 @@
|
||||
{
|
||||
"epoch": 0.9989528795811519,
|
||||
"total_flos": 0.0,
|
||||
"train_loss": 1.6700679651096433,
|
||||
"train_runtime": 3395.1223,
|
||||
"train_samples": 61135,
|
||||
"train_samples_per_second": 18.007,
|
||||
"train_steps_per_second": 0.14
|
||||
}
|
||||
795
trainer_state.json
Normal file
795
trainer_state.json
Normal file
@@ -0,0 +1,795 @@
|
||||
{
|
||||
"best_global_step": null,
|
||||
"best_metric": null,
|
||||
"best_model_checkpoint": null,
|
||||
"epoch": 0.9989528795811519,
|
||||
"eval_steps": 200,
|
||||
"global_step": 477,
|
||||
"is_hyper_param_search": false,
|
||||
"is_local_process_zero": true,
|
||||
"is_world_process_zero": true,
|
||||
"log_history": [
|
||||
{
|
||||
"epoch": 0.0020942408376963353,
|
||||
"grad_norm": 178.72232055664062,
|
||||
"learning_rate": 0.0,
|
||||
"logits/chosen": -0.615048885345459,
|
||||
"logits/rejected": -0.6184952855110168,
|
||||
"logps/chosen": -340.7542724609375,
|
||||
"logps/rejected": -224.71102905273438,
|
||||
"loss": 2.7815,
|
||||
"rewards/accuracies": 0.5625,
|
||||
"rewards/chosen": -0.0005076047964394093,
|
||||
"rewards/margins": 0.0013701794669032097,
|
||||
"rewards/rejected": -0.0018777847290039062,
|
||||
"step": 1
|
||||
},
|
||||
{
|
||||
"epoch": 0.020942408376963352,
|
||||
"grad_norm": 165.93289184570312,
|
||||
"learning_rate": 9.375e-08,
|
||||
"logits/chosen": -0.6438462734222412,
|
||||
"logits/rejected": -0.6377246379852295,
|
||||
"logps/chosen": -259.61663818359375,
|
||||
"logps/rejected": -258.05615234375,
|
||||
"loss": 2.7689,
|
||||
"rewards/accuracies": 0.5208333134651184,
|
||||
"rewards/chosen": 0.0028021351899951696,
|
||||
"rewards/margins": 0.0066328952088952065,
|
||||
"rewards/rejected": -0.0038307607173919678,
|
||||
"step": 10
|
||||
},
|
||||
{
|
||||
"epoch": 0.041884816753926704,
|
||||
"grad_norm": 185.72096252441406,
|
||||
"learning_rate": 1.9791666666666664e-07,
|
||||
"logits/chosen": -0.6060315370559692,
|
||||
"logits/rejected": -0.6294665336608887,
|
||||
"logps/chosen": -292.277587890625,
|
||||
"logps/rejected": -243.3096923828125,
|
||||
"loss": 2.7606,
|
||||
"rewards/accuracies": 0.5562499761581421,
|
||||
"rewards/chosen": 0.007751139812171459,
|
||||
"rewards/margins": 0.006523211486637592,
|
||||
"rewards/rejected": 0.001227927627041936,
|
||||
"step": 20
|
||||
},
|
||||
{
|
||||
"epoch": 0.06282722513089005,
|
||||
"grad_norm": 177.17030334472656,
|
||||
"learning_rate": 3.020833333333333e-07,
|
||||
"logits/chosen": -0.5830520987510681,
|
||||
"logits/rejected": -0.5622227191925049,
|
||||
"logps/chosen": -266.82086181640625,
|
||||
"logps/rejected": -252.55810546875,
|
||||
"loss": 2.7334,
|
||||
"rewards/accuracies": 0.5874999761581421,
|
||||
"rewards/chosen": 0.02364853397011757,
|
||||
"rewards/margins": 0.014692515134811401,
|
||||
"rewards/rejected": 0.008956017903983593,
|
||||
"step": 30
|
||||
},
|
||||
{
|
||||
"epoch": 0.08376963350785341,
|
||||
"grad_norm": 161.6173095703125,
|
||||
"learning_rate": 4.0625e-07,
|
||||
"logits/chosen": -0.610925555229187,
|
||||
"logits/rejected": -0.6422015428543091,
|
||||
"logps/chosen": -283.6108093261719,
|
||||
"logps/rejected": -263.230712890625,
|
||||
"loss": 2.6428,
|
||||
"rewards/accuracies": 0.65625,
|
||||
"rewards/chosen": 0.07614050805568695,
|
||||
"rewards/margins": 0.0455891415476799,
|
||||
"rewards/rejected": 0.030551373958587646,
|
||||
"step": 40
|
||||
},
|
||||
{
|
||||
"epoch": 0.10471204188481675,
|
||||
"grad_norm": 168.7168426513672,
|
||||
"learning_rate": 4.999932966293553e-07,
|
||||
"logits/chosen": -0.6716187596321106,
|
||||
"logits/rejected": -0.6886256337165833,
|
||||
"logps/chosen": -264.2910461425781,
|
||||
"logps/rejected": -256.81341552734375,
|
||||
"loss": 2.5166,
|
||||
"rewards/accuracies": 0.6875,
|
||||
"rewards/chosen": 0.14609374105930328,
|
||||
"rewards/margins": 0.11362485587596893,
|
||||
"rewards/rejected": 0.032468877732753754,
|
||||
"step": 50
|
||||
},
|
||||
{
|
||||
"epoch": 0.1256544502617801,
|
||||
"grad_norm": 161.653076171875,
|
||||
"learning_rate": 4.991893270335525e-07,
|
||||
"logits/chosen": -0.6373323798179626,
|
||||
"logits/rejected": -0.6400020718574524,
|
||||
"logps/chosen": -292.5552062988281,
|
||||
"logps/rejected": -283.8487243652344,
|
||||
"loss": 2.3448,
|
||||
"rewards/accuracies": 0.731249988079071,
|
||||
"rewards/chosen": 0.2640538215637207,
|
||||
"rewards/margins": 0.24639348685741425,
|
||||
"rewards/rejected": 0.01766030862927437,
|
||||
"step": 60
|
||||
},
|
||||
{
|
||||
"epoch": 0.14659685863874344,
|
||||
"grad_norm": 150.28529357910156,
|
||||
"learning_rate": 4.970496218214204e-07,
|
||||
"logits/chosen": -0.7395513653755188,
|
||||
"logits/rejected": -0.73731929063797,
|
||||
"logps/chosen": -294.8056945800781,
|
||||
"logps/rejected": -290.65618896484375,
|
||||
"loss": 2.2036,
|
||||
"rewards/accuracies": 0.706250011920929,
|
||||
"rewards/chosen": 0.38670551776885986,
|
||||
"rewards/margins": 0.33332276344299316,
|
||||
"rewards/rejected": 0.05338277295231819,
|
||||
"step": 70
|
||||
},
|
||||
{
|
||||
"epoch": 0.16753926701570682,
|
||||
"grad_norm": 150.14097595214844,
|
||||
"learning_rate": 4.935856505068998e-07,
|
||||
"logits/chosen": -0.6947239637374878,
|
||||
"logits/rejected": -0.7096772193908691,
|
||||
"logps/chosen": -265.2959289550781,
|
||||
"logps/rejected": -247.0114288330078,
|
||||
"loss": 2.0865,
|
||||
"rewards/accuracies": 0.675000011920929,
|
||||
"rewards/chosen": 0.44251909852027893,
|
||||
"rewards/margins": 0.39249858260154724,
|
||||
"rewards/rejected": 0.0500204861164093,
|
||||
"step": 80
|
||||
},
|
||||
{
|
||||
"epoch": 0.18848167539267016,
|
||||
"grad_norm": 158.30775451660156,
|
||||
"learning_rate": 4.8881598109976e-07,
|
||||
"logits/chosen": -0.6669536828994751,
|
||||
"logits/rejected": -0.699147641658783,
|
||||
"logps/chosen": -281.6817932128906,
|
||||
"logps/rejected": -246.4118194580078,
|
||||
"loss": 2.024,
|
||||
"rewards/accuracies": 0.7124999761581421,
|
||||
"rewards/chosen": 0.4167642593383789,
|
||||
"rewards/margins": 0.561462938785553,
|
||||
"rewards/rejected": -0.14469867944717407,
|
||||
"step": 90
|
||||
},
|
||||
{
|
||||
"epoch": 0.2094240837696335,
|
||||
"grad_norm": 159.54420471191406,
|
||||
"learning_rate": 4.827661805750437e-07,
|
||||
"logits/chosen": -0.6790419220924377,
|
||||
"logits/rejected": -0.6878204345703125,
|
||||
"logps/chosen": -258.7181701660156,
|
||||
"logps/rejected": -271.3045349121094,
|
||||
"loss": 2.0356,
|
||||
"rewards/accuracies": 0.6625000238418579,
|
||||
"rewards/chosen": 0.389530211687088,
|
||||
"rewards/margins": 0.5577359199523926,
|
||||
"rewards/rejected": -0.16820567846298218,
|
||||
"step": 100
|
||||
},
|
||||
{
|
||||
"epoch": 0.23036649214659685,
|
||||
"grad_norm": 148.32960510253906,
|
||||
"learning_rate": 4.75468677825789e-07,
|
||||
"logits/chosen": -0.649848461151123,
|
||||
"logits/rejected": -0.7013173699378967,
|
||||
"logps/chosen": -278.6050109863281,
|
||||
"logps/rejected": -287.51776123046875,
|
||||
"loss": 1.7628,
|
||||
"rewards/accuracies": 0.731249988079071,
|
||||
"rewards/chosen": 0.2824258506298065,
|
||||
"rewards/margins": 0.6535213589668274,
|
||||
"rewards/rejected": -0.3710955083370209,
|
||||
"step": 110
|
||||
},
|
||||
{
|
||||
"epoch": 0.2513089005235602,
|
||||
"grad_norm": 159.53060913085938,
|
||||
"learning_rate": 4.669625898336438e-07,
|
||||
"logits/chosen": -0.7270513772964478,
|
||||
"logits/rejected": -0.7322048544883728,
|
||||
"logps/chosen": -268.2157287597656,
|
||||
"logps/rejected": -264.9087829589844,
|
||||
"loss": 1.9148,
|
||||
"rewards/accuracies": 0.6937500238418579,
|
||||
"rewards/chosen": 0.10032379627227783,
|
||||
"rewards/margins": 0.6243712902069092,
|
||||
"rewards/rejected": -0.5240474343299866,
|
||||
"step": 120
|
||||
},
|
||||
{
|
||||
"epoch": 0.27225130890052357,
|
||||
"grad_norm": 180.24705505371094,
|
||||
"learning_rate": 4.5729351198915705e-07,
|
||||
"logits/chosen": -0.7340003848075867,
|
||||
"logits/rejected": -0.718673825263977,
|
||||
"logps/chosen": -288.23260498046875,
|
||||
"logps/rejected": -291.12969970703125,
|
||||
"loss": 1.8436,
|
||||
"rewards/accuracies": 0.6312500238418579,
|
||||
"rewards/chosen": 0.08194615691900253,
|
||||
"rewards/margins": 0.6275433897972107,
|
||||
"rewards/rejected": -0.5455971956253052,
|
||||
"step": 130
|
||||
},
|
||||
{
|
||||
"epoch": 0.2931937172774869,
|
||||
"grad_norm": 163.4424285888672,
|
||||
"learning_rate": 4.4651327368569684e-07,
|
||||
"logits/chosen": -0.7740471363067627,
|
||||
"logits/rejected": -0.7457703351974487,
|
||||
"logps/chosen": -292.3294982910156,
|
||||
"logps/rejected": -274.18505859375,
|
||||
"loss": 1.8836,
|
||||
"rewards/accuracies": 0.6499999761581421,
|
||||
"rewards/chosen": 0.17513397336006165,
|
||||
"rewards/margins": 0.7150871157646179,
|
||||
"rewards/rejected": -0.5399531126022339,
|
||||
"step": 140
|
||||
},
|
||||
{
|
||||
"epoch": 0.31413612565445026,
|
||||
"grad_norm": 183.37110900878906,
|
||||
"learning_rate": 4.346796604970912e-07,
|
||||
"logits/chosen": -0.7887325882911682,
|
||||
"logits/rejected": -0.7826852202415466,
|
||||
"logps/chosen": -306.27752685546875,
|
||||
"logps/rejected": -263.0813903808594,
|
||||
"loss": 1.6791,
|
||||
"rewards/accuracies": 0.7250000238418579,
|
||||
"rewards/chosen": 0.32484668493270874,
|
||||
"rewards/margins": 1.0725610256195068,
|
||||
"rewards/rejected": -0.7477144002914429,
|
||||
"step": 150
|
||||
},
|
||||
{
|
||||
"epoch": 0.33507853403141363,
|
||||
"grad_norm": 154.94068908691406,
|
||||
"learning_rate": 4.218561044282098e-07,
|
||||
"logits/chosen": -0.7592242956161499,
|
||||
"logits/rejected": -0.7694119215011597,
|
||||
"logps/chosen": -269.1500244140625,
|
||||
"logps/rejected": -244.9143829345703,
|
||||
"loss": 1.4581,
|
||||
"rewards/accuracies": 0.793749988079071,
|
||||
"rewards/chosen": -0.11840873956680298,
|
||||
"rewards/margins": 1.2886579036712646,
|
||||
"rewards/rejected": -1.4070665836334229,
|
||||
"step": 160
|
||||
},
|
||||
{
|
||||
"epoch": 0.35602094240837695,
|
||||
"grad_norm": 188.69107055664062,
|
||||
"learning_rate": 4.081113438988443e-07,
|
||||
"logits/chosen": -0.7660834193229675,
|
||||
"logits/rejected": -0.7688754796981812,
|
||||
"logps/chosen": -288.95806884765625,
|
||||
"logps/rejected": -266.93170166015625,
|
||||
"loss": 1.5308,
|
||||
"rewards/accuracies": 0.75,
|
||||
"rewards/chosen": 0.014453152194619179,
|
||||
"rewards/margins": 1.154816746711731,
|
||||
"rewards/rejected": -1.1403634548187256,
|
||||
"step": 170
|
||||
},
|
||||
{
|
||||
"epoch": 0.3769633507853403,
|
||||
"grad_norm": 145.07247924804688,
|
||||
"learning_rate": 3.935190552834828e-07,
|
||||
"logits/chosen": -0.7244433760643005,
|
||||
"logits/rejected": -0.7473747134208679,
|
||||
"logps/chosen": -271.08758544921875,
|
||||
"logps/rejected": -252.1222686767578,
|
||||
"loss": 1.6745,
|
||||
"rewards/accuracies": 0.731249988079071,
|
||||
"rewards/chosen": -0.007142066955566406,
|
||||
"rewards/margins": 1.086653470993042,
|
||||
"rewards/rejected": -1.093795657157898,
|
||||
"step": 180
|
||||
},
|
||||
{
|
||||
"epoch": 0.39790575916230364,
|
||||
"grad_norm": 160.0677490234375,
|
||||
"learning_rate": 3.781574579820464e-07,
|
||||
"logits/chosen": -0.7775353193283081,
|
||||
"logits/rejected": -0.7835763692855835,
|
||||
"logps/chosen": -284.01068115234375,
|
||||
"logps/rejected": -260.91009521484375,
|
||||
"loss": 1.4346,
|
||||
"rewards/accuracies": 0.7562500238418579,
|
||||
"rewards/chosen": 0.1680121123790741,
|
||||
"rewards/margins": 1.4267839193344116,
|
||||
"rewards/rejected": -1.2587717771530151,
|
||||
"step": 190
|
||||
},
|
||||
{
|
||||
"epoch": 0.418848167539267,
|
||||
"grad_norm": 195.3789520263672,
|
||||
"learning_rate": 3.621088951385353e-07,
|
||||
"logits/chosen": -0.7822957634925842,
|
||||
"logits/rejected": -0.8039867281913757,
|
||||
"logps/chosen": -267.6298522949219,
|
||||
"logps/rejected": -259.41973876953125,
|
||||
"loss": 1.5682,
|
||||
"rewards/accuracies": 0.6875,
|
||||
"rewards/chosen": 0.25724172592163086,
|
||||
"rewards/margins": 1.3147997856140137,
|
||||
"rewards/rejected": -1.0575581789016724,
|
||||
"step": 200
|
||||
},
|
||||
{
|
||||
"epoch": 0.418848167539267,
|
||||
"eval_logits/chosen": -0.800130307674408,
|
||||
"eval_logits/rejected": -0.7831319570541382,
|
||||
"eval_logps/chosen": -300.2963562011719,
|
||||
"eval_logps/rejected": -282.6593322753906,
|
||||
"eval_loss": 0.4018401503562927,
|
||||
"eval_rewards/accuracies": 0.7540322542190552,
|
||||
"eval_rewards/chosen": -0.24218061566352844,
|
||||
"eval_rewards/margins": 1.2888014316558838,
|
||||
"eval_rewards/rejected": -1.5309820175170898,
|
||||
"eval_runtime": 44.8688,
|
||||
"eval_samples_per_second": 44.574,
|
||||
"eval_steps_per_second": 1.404,
|
||||
"step": 200
|
||||
},
|
||||
{
|
||||
"epoch": 0.4397905759162304,
|
||||
"grad_norm": 197.96990966796875,
|
||||
"learning_rate": 3.454593922550693e-07,
|
||||
"logits/chosen": -0.7248128056526184,
|
||||
"logits/rejected": -0.7440636157989502,
|
||||
"logps/chosen": -288.216552734375,
|
||||
"logps/rejected": -284.04571533203125,
|
||||
"loss": 1.5993,
|
||||
"rewards/accuracies": 0.78125,
|
||||
"rewards/chosen": -0.09156915545463562,
|
||||
"rewards/margins": 1.5361555814743042,
|
||||
"rewards/rejected": -1.6277246475219727,
|
||||
"step": 210
|
||||
},
|
||||
{
|
||||
"epoch": 0.4607329842931937,
|
||||
"grad_norm": 184.7930145263672,
|
||||
"learning_rate": 3.2829819606729477e-07,
|
||||
"logits/chosen": -0.7869287133216858,
|
||||
"logits/rejected": -0.770258367061615,
|
||||
"logps/chosen": -275.52459716796875,
|
||||
"logps/rejected": -306.2828063964844,
|
||||
"loss": 1.3904,
|
||||
"rewards/accuracies": 0.75,
|
||||
"rewards/chosen": -0.46098509430885315,
|
||||
"rewards/margins": 1.4248541593551636,
|
||||
"rewards/rejected": -1.8858392238616943,
|
||||
"step": 220
|
||||
},
|
||||
{
|
||||
"epoch": 0.4816753926701571,
|
||||
"grad_norm": 209.72683715820312,
|
||||
"learning_rate": 3.1071729615293424e-07,
|
||||
"logits/chosen": -0.7838747501373291,
|
||||
"logits/rejected": -0.7623480558395386,
|
||||
"logps/chosen": -318.6150207519531,
|
||||
"logps/rejected": -280.46673583984375,
|
||||
"loss": 1.2821,
|
||||
"rewards/accuracies": 0.793749988079071,
|
||||
"rewards/chosen": -0.25955599546432495,
|
||||
"rewards/margins": 1.878294587135315,
|
||||
"rewards/rejected": -2.137850761413574,
|
||||
"step": 230
|
||||
},
|
||||
{
|
||||
"epoch": 0.5026178010471204,
|
||||
"grad_norm": 189.98550415039062,
|
||||
"learning_rate": 2.9281093183781403e-07,
|
||||
"logits/chosen": -0.7183490991592407,
|
||||
"logits/rejected": -0.7486238479614258,
|
||||
"logps/chosen": -254.42953491210938,
|
||||
"logps/rejected": -258.20489501953125,
|
||||
"loss": 1.492,
|
||||
"rewards/accuracies": 0.7250000238418579,
|
||||
"rewards/chosen": -0.5049279928207397,
|
||||
"rewards/margins": 1.325303077697754,
|
||||
"rewards/rejected": -1.8302310705184937,
|
||||
"step": 240
|
||||
},
|
||||
{
|
||||
"epoch": 0.5235602094240838,
|
||||
"grad_norm": 206.22911071777344,
|
||||
"learning_rate": 2.7467508704251135e-07,
|
||||
"logits/chosen": -0.7619749903678894,
|
||||
"logits/rejected": -0.7551571726799011,
|
||||
"logps/chosen": -296.3054504394531,
|
||||
"logps/rejected": -302.1717224121094,
|
||||
"loss": 1.5437,
|
||||
"rewards/accuracies": 0.7562500238418579,
|
||||
"rewards/chosen": -0.023396695032715797,
|
||||
"rewards/margins": 1.536604404449463,
|
||||
"rewards/rejected": -1.5600011348724365,
|
||||
"step": 250
|
||||
},
|
||||
{
|
||||
"epoch": 0.5445026178010471,
|
||||
"grad_norm": 225.17259216308594,
|
||||
"learning_rate": 2.5640697577740815e-07,
|
||||
"logits/chosen": -0.7599934935569763,
|
||||
"logits/rejected": -0.760217547416687,
|
||||
"logps/chosen": -283.0222473144531,
|
||||
"logps/rejected": -279.87725830078125,
|
||||
"loss": 1.4859,
|
||||
"rewards/accuracies": 0.7749999761581421,
|
||||
"rewards/chosen": -0.262519896030426,
|
||||
"rewards/margins": 1.3884146213531494,
|
||||
"rewards/rejected": -1.6509345769882202,
|
||||
"step": 260
|
||||
},
|
||||
{
|
||||
"epoch": 0.5654450261780105,
|
||||
"grad_norm": 178.20997619628906,
|
||||
"learning_rate": 2.381045210440644e-07,
|
||||
"logits/chosen": -0.751593828201294,
|
||||
"logits/rejected": -0.740311861038208,
|
||||
"logps/chosen": -262.66802978515625,
|
||||
"logps/rejected": -278.4461975097656,
|
||||
"loss": 1.3796,
|
||||
"rewards/accuracies": 0.75,
|
||||
"rewards/chosen": -0.4121341109275818,
|
||||
"rewards/margins": 1.4869859218597412,
|
||||
"rewards/rejected": -1.8991199731826782,
|
||||
"step": 270
|
||||
},
|
||||
{
|
||||
"epoch": 0.5863874345549738,
|
||||
"grad_norm": 166.0771026611328,
|
||||
"learning_rate": 2.1986582993616925e-07,
|
||||
"logits/chosen": -0.7442865967750549,
|
||||
"logits/rejected": -0.7617285847663879,
|
||||
"logps/chosen": -291.0567321777344,
|
||||
"logps/rejected": -286.50579833984375,
|
||||
"loss": 1.3057,
|
||||
"rewards/accuracies": 0.737500011920929,
|
||||
"rewards/chosen": -0.2755025029182434,
|
||||
"rewards/margins": 1.6462242603302002,
|
||||
"rewards/rejected": -1.9217264652252197,
|
||||
"step": 280
|
||||
},
|
||||
{
|
||||
"epoch": 0.6073298429319371,
|
||||
"grad_norm": 179.26734924316406,
|
||||
"learning_rate": 2.0178866775369774e-07,
|
||||
"logits/chosen": -0.7708907723426819,
|
||||
"logits/rejected": -0.7562252879142761,
|
||||
"logps/chosen": -316.54803466796875,
|
||||
"logps/rejected": -304.83404541015625,
|
||||
"loss": 1.5853,
|
||||
"rewards/accuracies": 0.7124999761581421,
|
||||
"rewards/chosen": -0.6070529818534851,
|
||||
"rewards/margins": 1.3637443780899048,
|
||||
"rewards/rejected": -1.9707973003387451,
|
||||
"step": 290
|
||||
},
|
||||
{
|
||||
"epoch": 0.6282722513089005,
|
||||
"grad_norm": 188.23605346679688,
|
||||
"learning_rate": 1.839699339491937e-07,
|
||||
"logits/chosen": -0.7966352105140686,
|
||||
"logits/rejected": -0.7720645666122437,
|
||||
"logps/chosen": -269.08026123046875,
|
||||
"logps/rejected": -288.74798583984375,
|
||||
"loss": 1.5072,
|
||||
"rewards/accuracies": 0.731249988079071,
|
||||
"rewards/chosen": 0.051731567829847336,
|
||||
"rewards/margins": 1.4731868505477905,
|
||||
"rewards/rejected": -1.4214551448822021,
|
||||
"step": 300
|
||||
},
|
||||
{
|
||||
"epoch": 0.6492146596858639,
|
||||
"grad_norm": 193.12405395507812,
|
||||
"learning_rate": 1.6650514271527465e-07,
|
||||
"logits/chosen": -0.7465989589691162,
|
||||
"logits/rejected": -0.732526421546936,
|
||||
"logps/chosen": -315.72998046875,
|
||||
"logps/rejected": -311.20135498046875,
|
||||
"loss": 1.3191,
|
||||
"rewards/accuracies": 0.7749999761581421,
|
||||
"rewards/chosen": 0.06771044433116913,
|
||||
"rewards/margins": 1.4948314428329468,
|
||||
"rewards/rejected": -1.4271211624145508,
|
||||
"step": 310
|
||||
},
|
||||
{
|
||||
"epoch": 0.6701570680628273,
|
||||
"grad_norm": 208.85194396972656,
|
||||
"learning_rate": 1.4948791099758052e-07,
|
||||
"logits/chosen": -0.7754351496696472,
|
||||
"logits/rejected": -0.7965367436408997,
|
||||
"logps/chosen": -287.69244384765625,
|
||||
"logps/rejected": -258.49078369140625,
|
||||
"loss": 1.3408,
|
||||
"rewards/accuracies": 0.7250000238418579,
|
||||
"rewards/chosen": -0.2836495339870453,
|
||||
"rewards/margins": 1.5221548080444336,
|
||||
"rewards/rejected": -1.8058042526245117,
|
||||
"step": 320
|
||||
},
|
||||
{
|
||||
"epoch": 0.6910994764397905,
|
||||
"grad_norm": 196.4761962890625,
|
||||
"learning_rate": 1.3300945667758012e-07,
|
||||
"logits/chosen": -0.7553143501281738,
|
||||
"logits/rejected": -0.7459635138511658,
|
||||
"logps/chosen": -284.33258056640625,
|
||||
"logps/rejected": -292.96905517578125,
|
||||
"loss": 1.3833,
|
||||
"rewards/accuracies": 0.8125,
|
||||
"rewards/chosen": -0.4384119510650635,
|
||||
"rewards/margins": 1.677755355834961,
|
||||
"rewards/rejected": -2.1161673069000244,
|
||||
"step": 330
|
||||
},
|
||||
{
|
||||
"epoch": 0.7120418848167539,
|
||||
"grad_norm": 196.71824645996094,
|
||||
"learning_rate": 1.1715810961514072e-07,
|
||||
"logits/chosen": -0.7940613031387329,
|
||||
"logits/rejected": -0.7792786955833435,
|
||||
"logps/chosen": -260.6127624511719,
|
||||
"logps/rejected": -280.0231018066406,
|
||||
"loss": 1.3467,
|
||||
"rewards/accuracies": 0.7124999761581421,
|
||||
"rewards/chosen": -0.8941448330879211,
|
||||
"rewards/margins": 1.461021065711975,
|
||||
"rewards/rejected": -2.355165958404541,
|
||||
"step": 340
|
||||
},
|
||||
{
|
||||
"epoch": 0.7329842931937173,
|
||||
"grad_norm": 206.063232421875,
|
||||
"learning_rate": 1.0201883817182949e-07,
|
||||
"logits/chosen": -0.8100167512893677,
|
||||
"logits/rejected": -0.8022691011428833,
|
||||
"logps/chosen": -272.3794250488281,
|
||||
"logps/rejected": -279.54290771484375,
|
||||
"loss": 1.4061,
|
||||
"rewards/accuracies": 0.706250011920929,
|
||||
"rewards/chosen": -0.7607845664024353,
|
||||
"rewards/margins": 1.3687453269958496,
|
||||
"rewards/rejected": -2.1295297145843506,
|
||||
"step": 350
|
||||
},
|
||||
{
|
||||
"epoch": 0.7539267015706806,
|
||||
"grad_norm": 180.1078338623047,
|
||||
"learning_rate": 8.76727937529367e-08,
|
||||
"logits/chosen": -0.7735427021980286,
|
||||
"logits/rejected": -0.7678244709968567,
|
||||
"logps/chosen": -281.62591552734375,
|
||||
"logps/rejected": -278.86700439453125,
|
||||
"loss": 1.5415,
|
||||
"rewards/accuracies": 0.706250011920929,
|
||||
"rewards/chosen": -0.6388282775878906,
|
||||
"rewards/margins": 1.561563491821289,
|
||||
"rewards/rejected": -2.2003917694091797,
|
||||
"step": 360
|
||||
},
|
||||
{
|
||||
"epoch": 0.774869109947644,
|
||||
"grad_norm": 159.87562561035156,
|
||||
"learning_rate": 7.419687580962222e-08,
|
||||
"logits/chosen": -0.7762534618377686,
|
||||
"logits/rejected": -0.7863459587097168,
|
||||
"logps/chosen": -306.4834289550781,
|
||||
"logps/rejected": -304.45135498046875,
|
||||
"loss": 1.1556,
|
||||
"rewards/accuracies": 0.78125,
|
||||
"rewards/chosen": -0.3031012713909149,
|
||||
"rewards/margins": 1.8383687734603882,
|
||||
"rewards/rejected": -2.141469955444336,
|
||||
"step": 370
|
||||
},
|
||||
{
|
||||
"epoch": 0.7958115183246073,
|
||||
"grad_norm": 225.36546325683594,
|
||||
"learning_rate": 6.166331963291519e-08,
|
||||
"logits/chosen": -0.7447551488876343,
|
||||
"logits/rejected": -0.7403131723403931,
|
||||
"logps/chosen": -300.09930419921875,
|
||||
"logps/rejected": -274.04351806640625,
|
||||
"loss": 1.4064,
|
||||
"rewards/accuracies": 0.78125,
|
||||
"rewards/chosen": -0.3262189030647278,
|
||||
"rewards/margins": 1.6647300720214844,
|
||||
"rewards/rejected": -1.9909489154815674,
|
||||
"step": 380
|
||||
},
|
||||
{
|
||||
"epoch": 0.8167539267015707,
|
||||
"grad_norm": 182.28985595703125,
|
||||
"learning_rate": 5.013930914912476e-08,
|
||||
"logits/chosen": -0.8010516166687012,
|
||||
"logits/rejected": -0.7907552719116211,
|
||||
"logps/chosen": -266.9259338378906,
|
||||
"logps/rejected": -266.32281494140625,
|
||||
"loss": 1.257,
|
||||
"rewards/accuracies": 0.7437499761581421,
|
||||
"rewards/chosen": -0.2791803777217865,
|
||||
"rewards/margins": 1.4971604347229004,
|
||||
"rewards/rejected": -1.7763407230377197,
|
||||
"step": 390
|
||||
},
|
||||
{
|
||||
"epoch": 0.837696335078534,
|
||||
"grad_norm": 180.9009246826172,
|
||||
"learning_rate": 3.968661679220467e-08,
|
||||
"logits/chosen": -0.7620838284492493,
|
||||
"logits/rejected": -0.7844998240470886,
|
||||
"logps/chosen": -293.094482421875,
|
||||
"logps/rejected": -261.71636962890625,
|
||||
"loss": 1.3853,
|
||||
"rewards/accuracies": 0.731249988079071,
|
||||
"rewards/chosen": -0.32791513204574585,
|
||||
"rewards/margins": 1.4900782108306885,
|
||||
"rewards/rejected": -1.8179931640625,
|
||||
"step": 400
|
||||
},
|
||||
{
|
||||
"epoch": 0.837696335078534,
|
||||
"eval_logits/chosen": -0.7842384576797485,
|
||||
"eval_logits/rejected": -0.7659946084022522,
|
||||
"eval_logps/chosen": -300.5436706542969,
|
||||
"eval_logps/rejected": -286.3951416015625,
|
||||
"eval_loss": 0.3504386246204376,
|
||||
"eval_rewards/accuracies": 0.7459677457809448,
|
||||
"eval_rewards/chosen": -0.2669134736061096,
|
||||
"eval_rewards/margins": 1.6376467943191528,
|
||||
"eval_rewards/rejected": -1.9045602083206177,
|
||||
"eval_runtime": 44.6161,
|
||||
"eval_samples_per_second": 44.827,
|
||||
"eval_steps_per_second": 1.412,
|
||||
"step": 400
|
||||
},
|
||||
{
|
||||
"epoch": 0.8586387434554974,
|
||||
"grad_norm": 201.3262176513672,
|
||||
"learning_rate": 3.036127238347164e-08,
|
||||
"logits/chosen": -0.7589632272720337,
|
||||
"logits/rejected": -0.7365776896476746,
|
||||
"logps/chosen": -272.2435302734375,
|
||||
"logps/rejected": -303.0001525878906,
|
||||
"loss": 1.3037,
|
||||
"rewards/accuracies": 0.6937500238418579,
|
||||
"rewards/chosen": -0.48334985971450806,
|
||||
"rewards/margins": 1.4791133403778076,
|
||||
"rewards/rejected": -1.9624630212783813,
|
||||
"step": 410
|
||||
},
|
||||
{
|
||||
"epoch": 0.8795811518324608,
|
||||
"grad_norm": 210.60845947265625,
|
||||
"learning_rate": 2.2213262793589482e-08,
|
||||
"logits/chosen": -0.7427655458450317,
|
||||
"logits/rejected": -0.733137309551239,
|
||||
"logps/chosen": -323.9688415527344,
|
||||
"logps/rejected": -285.21478271484375,
|
||||
"loss": 1.236,
|
||||
"rewards/accuracies": 0.6812499761581421,
|
||||
"rewards/chosen": -0.3501654267311096,
|
||||
"rewards/margins": 1.6594756841659546,
|
||||
"rewards/rejected": -2.009641170501709,
|
||||
"step": 420
|
||||
},
|
||||
{
|
||||
"epoch": 0.900523560209424,
|
||||
"grad_norm": 193.39053344726562,
|
||||
"learning_rate": 1.5286263996730026e-08,
|
||||
"logits/chosen": -0.7872425317764282,
|
||||
"logits/rejected": -0.7826786041259766,
|
||||
"logps/chosen": -306.154052734375,
|
||||
"logps/rejected": -300.08172607421875,
|
||||
"loss": 1.2994,
|
||||
"rewards/accuracies": 0.71875,
|
||||
"rewards/chosen": -0.4893164038658142,
|
||||
"rewards/margins": 1.622433066368103,
|
||||
"rewards/rejected": -2.1117494106292725,
|
||||
"step": 430
|
||||
},
|
||||
{
|
||||
"epoch": 0.9214659685863874,
|
||||
"grad_norm": 236.22662353515625,
|
||||
"learning_rate": 9.617406953185136e-09,
|
||||
"logits/chosen": -0.7628771662712097,
|
||||
"logits/rejected": -0.7661261558532715,
|
||||
"logps/chosen": -288.91278076171875,
|
||||
"logps/rejected": -315.32373046875,
|
||||
"loss": 1.5465,
|
||||
"rewards/accuracies": 0.6625000238418579,
|
||||
"rewards/chosen": -0.37518563866615295,
|
||||
"rewards/margins": 1.1774741411209106,
|
||||
"rewards/rejected": -1.5526599884033203,
|
||||
"step": 440
|
||||
},
|
||||
{
|
||||
"epoch": 0.9424083769633508,
|
||||
"grad_norm": 187.93885803222656,
|
||||
"learning_rate": 5.2370785753763356e-09,
|
||||
"logits/chosen": -0.7725846767425537,
|
||||
"logits/rejected": -0.7734043002128601,
|
||||
"logps/chosen": -284.969970703125,
|
||||
"logps/rejected": -260.22076416015625,
|
||||
"loss": 1.1488,
|
||||
"rewards/accuracies": 0.7562500238418579,
|
||||
"rewards/chosen": -0.275113046169281,
|
||||
"rewards/margins": 1.5526530742645264,
|
||||
"rewards/rejected": -1.8277660608291626,
|
||||
"step": 450
|
||||
},
|
||||
{
|
||||
"epoch": 0.9633507853403142,
|
||||
"grad_norm": 191.54318237304688,
|
||||
"learning_rate": 2.168758844148272e-09,
|
||||
"logits/chosen": -0.7661501169204712,
|
||||
"logits/rejected": -0.7916361689567566,
|
||||
"logps/chosen": -326.5988464355469,
|
||||
"logps/rejected": -314.1842041015625,
|
||||
"loss": 1.485,
|
||||
"rewards/accuracies": 0.71875,
|
||||
"rewards/chosen": -0.322204053401947,
|
||||
"rewards/margins": 1.6917155981063843,
|
||||
"rewards/rejected": -2.0139195919036865,
|
||||
"step": 460
|
||||
},
|
||||
{
|
||||
"epoch": 0.9842931937172775,
|
||||
"grad_norm": 176.15220642089844,
|
||||
"learning_rate": 4.288949484559934e-10,
|
||||
"logits/chosen": -0.7752399444580078,
|
||||
"logits/rejected": -0.7753234505653381,
|
||||
"logps/chosen": -258.8638916015625,
|
||||
"logps/rejected": -276.8975524902344,
|
||||
"loss": 1.2765,
|
||||
"rewards/accuracies": 0.7562500238418579,
|
||||
"rewards/chosen": -0.06555650383234024,
|
||||
"rewards/margins": 1.781582236289978,
|
||||
"rewards/rejected": -1.8471390008926392,
|
||||
"step": 470
|
||||
},
|
||||
{
|
||||
"epoch": 0.9989528795811519,
|
||||
"step": 477,
|
||||
"total_flos": 0.0,
|
||||
"train_loss": 1.6700679651096433,
|
||||
"train_runtime": 3395.1223,
|
||||
"train_samples_per_second": 18.007,
|
||||
"train_steps_per_second": 0.14
|
||||
}
|
||||
],
|
||||
"logging_steps": 10,
|
||||
"max_steps": 477,
|
||||
"num_input_tokens_seen": 0,
|
||||
"num_train_epochs": 1,
|
||||
"save_steps": 500,
|
||||
"stateful_callbacks": {
|
||||
"TrainerControl": {
|
||||
"args": {
|
||||
"should_epoch_stop": false,
|
||||
"should_evaluate": false,
|
||||
"should_log": false,
|
||||
"should_save": false,
|
||||
"should_training_stop": false
|
||||
},
|
||||
"attributes": {}
|
||||
}
|
||||
},
|
||||
"total_flos": 0.0,
|
||||
"train_batch_size": 4,
|
||||
"trial_name": null,
|
||||
"trial_params": null
|
||||
}
|
||||
Reference in New Issue
Block a user