Files
NTK_64k_Slimpajama_2B/trainer_state.json
ModelHub XC ce5041bbfc 初始化项目,由ModelHub XC社区提供模型
Model: Leooyii/NTK_64k_Slimpajama_2B
Source: Original Platform
2026-04-11 14:40:00 +08:00

749 lines
14 KiB
JSON

{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.997920997920998,
"eval_steps": 500,
"global_step": 120,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01,
"learning_rate": 0.0,
"loss": 5.3569,
"step": 1
},
{
"epoch": 0.02,
"learning_rate": 4.6275642631951835e-06,
"loss": 4.4928,
"step": 2
},
{
"epoch": 0.02,
"learning_rate": 7.3345158268416935e-06,
"loss": 4.4425,
"step": 3
},
{
"epoch": 0.03,
"learning_rate": 9.255128526390367e-06,
"loss": 3.7693,
"step": 4
},
{
"epoch": 0.04,
"learning_rate": 1.0744871473609633e-05,
"loss": 3.3417,
"step": 5
},
{
"epoch": 0.05,
"learning_rate": 1.1962080090036879e-05,
"loss": 2.7971,
"step": 6
},
{
"epoch": 0.06,
"learning_rate": 1.299121531141887e-05,
"loss": 2.5958,
"step": 7
},
{
"epoch": 0.07,
"learning_rate": 1.388269278958555e-05,
"loss": 2.383,
"step": 8
},
{
"epoch": 0.07,
"learning_rate": 1.4669031653683387e-05,
"loss": 2.1741,
"step": 9
},
{
"epoch": 0.08,
"learning_rate": 1.537243573680482e-05,
"loss": 2.0406,
"step": 10
},
{
"epoch": 0.09,
"learning_rate": 1.600874212937343e-05,
"loss": 1.9426,
"step": 11
},
{
"epoch": 0.1,
"learning_rate": 1.6589644353232063e-05,
"loss": 1.9206,
"step": 12
},
{
"epoch": 0.11,
"learning_rate": 1.712402259777778e-05,
"loss": 1.839,
"step": 13
},
{
"epoch": 0.12,
"learning_rate": 1.7618779574614054e-05,
"loss": 1.8111,
"step": 14
},
{
"epoch": 0.12,
"learning_rate": 1.8079387300451327e-05,
"loss": 1.7821,
"step": 15
},
{
"epoch": 0.13,
"learning_rate": 1.8510257052780734e-05,
"loss": 1.7383,
"step": 16
},
{
"epoch": 0.14,
"learning_rate": 1.891499697130832e-05,
"loss": 1.7587,
"step": 17
},
{
"epoch": 0.15,
"learning_rate": 1.929659591687857e-05,
"loss": 1.6917,
"step": 18
},
{
"epoch": 0.16,
"learning_rate": 1.9657557553855117e-05,
"loss": 1.7491,
"step": 19
},
{
"epoch": 0.17,
"learning_rate": 2e-05,
"loss": 1.677,
"step": 20
},
{
"epoch": 0.17,
"learning_rate": 2e-05,
"loss": 1.6952,
"step": 21
},
{
"epoch": 0.18,
"learning_rate": 1.98e-05,
"loss": 1.6802,
"step": 22
},
{
"epoch": 0.19,
"learning_rate": 1.9600000000000002e-05,
"loss": 1.6731,
"step": 23
},
{
"epoch": 0.2,
"learning_rate": 1.94e-05,
"loss": 1.6718,
"step": 24
},
{
"epoch": 0.21,
"learning_rate": 1.9200000000000003e-05,
"loss": 1.6927,
"step": 25
},
{
"epoch": 0.22,
"learning_rate": 1.9e-05,
"loss": 1.6631,
"step": 26
},
{
"epoch": 0.22,
"learning_rate": 1.88e-05,
"loss": 1.6413,
"step": 27
},
{
"epoch": 0.23,
"learning_rate": 1.86e-05,
"loss": 1.6799,
"step": 28
},
{
"epoch": 0.24,
"learning_rate": 1.8400000000000003e-05,
"loss": 1.6413,
"step": 29
},
{
"epoch": 0.25,
"learning_rate": 1.8200000000000002e-05,
"loss": 1.6549,
"step": 30
},
{
"epoch": 0.26,
"learning_rate": 1.8e-05,
"loss": 1.6693,
"step": 31
},
{
"epoch": 0.27,
"learning_rate": 1.7800000000000002e-05,
"loss": 1.667,
"step": 32
},
{
"epoch": 0.27,
"learning_rate": 1.76e-05,
"loss": 1.6308,
"step": 33
},
{
"epoch": 0.28,
"learning_rate": 1.7400000000000003e-05,
"loss": 1.6509,
"step": 34
},
{
"epoch": 0.29,
"learning_rate": 1.72e-05,
"loss": 1.6222,
"step": 35
},
{
"epoch": 0.3,
"learning_rate": 1.7e-05,
"loss": 1.6285,
"step": 36
},
{
"epoch": 0.31,
"learning_rate": 1.6800000000000002e-05,
"loss": 1.6167,
"step": 37
},
{
"epoch": 0.32,
"learning_rate": 1.66e-05,
"loss": 1.6296,
"step": 38
},
{
"epoch": 0.32,
"learning_rate": 1.64e-05,
"loss": 1.6444,
"step": 39
},
{
"epoch": 0.33,
"learning_rate": 1.62e-05,
"loss": 1.6628,
"step": 40
},
{
"epoch": 0.34,
"learning_rate": 1.6000000000000003e-05,
"loss": 1.6141,
"step": 41
},
{
"epoch": 0.35,
"learning_rate": 1.58e-05,
"loss": 1.639,
"step": 42
},
{
"epoch": 0.36,
"learning_rate": 1.5600000000000003e-05,
"loss": 1.618,
"step": 43
},
{
"epoch": 0.37,
"learning_rate": 1.54e-05,
"loss": 1.5938,
"step": 44
},
{
"epoch": 0.37,
"learning_rate": 1.5200000000000002e-05,
"loss": 1.6398,
"step": 45
},
{
"epoch": 0.38,
"learning_rate": 1.5000000000000002e-05,
"loss": 1.6212,
"step": 46
},
{
"epoch": 0.39,
"learning_rate": 1.48e-05,
"loss": 1.6495,
"step": 47
},
{
"epoch": 0.4,
"learning_rate": 1.46e-05,
"loss": 1.6114,
"step": 48
},
{
"epoch": 0.41,
"learning_rate": 1.4400000000000001e-05,
"loss": 1.643,
"step": 49
},
{
"epoch": 0.42,
"learning_rate": 1.4200000000000001e-05,
"loss": 1.6174,
"step": 50
},
{
"epoch": 0.42,
"learning_rate": 1.4e-05,
"loss": 1.5991,
"step": 51
},
{
"epoch": 0.43,
"learning_rate": 1.38e-05,
"loss": 1.6355,
"step": 52
},
{
"epoch": 0.44,
"learning_rate": 1.3600000000000002e-05,
"loss": 1.6439,
"step": 53
},
{
"epoch": 0.45,
"learning_rate": 1.3400000000000002e-05,
"loss": 1.6249,
"step": 54
},
{
"epoch": 0.46,
"learning_rate": 1.3200000000000002e-05,
"loss": 1.6543,
"step": 55
},
{
"epoch": 0.47,
"learning_rate": 1.3000000000000001e-05,
"loss": 1.6403,
"step": 56
},
{
"epoch": 0.47,
"learning_rate": 1.2800000000000001e-05,
"loss": 1.6351,
"step": 57
},
{
"epoch": 0.48,
"learning_rate": 1.2600000000000001e-05,
"loss": 1.6226,
"step": 58
},
{
"epoch": 0.49,
"learning_rate": 1.2400000000000002e-05,
"loss": 1.6459,
"step": 59
},
{
"epoch": 0.5,
"learning_rate": 1.22e-05,
"loss": 1.5992,
"step": 60
},
{
"epoch": 0.51,
"learning_rate": 1.2e-05,
"loss": 1.6293,
"step": 61
},
{
"epoch": 0.52,
"learning_rate": 1.18e-05,
"loss": 1.6327,
"step": 62
},
{
"epoch": 0.52,
"learning_rate": 1.16e-05,
"loss": 1.6132,
"step": 63
},
{
"epoch": 0.53,
"learning_rate": 1.14e-05,
"loss": 1.6111,
"step": 64
},
{
"epoch": 0.54,
"learning_rate": 1.1200000000000001e-05,
"loss": 1.6142,
"step": 65
},
{
"epoch": 0.55,
"learning_rate": 1.1000000000000001e-05,
"loss": 1.6114,
"step": 66
},
{
"epoch": 0.56,
"learning_rate": 1.0800000000000002e-05,
"loss": 1.6308,
"step": 67
},
{
"epoch": 0.57,
"learning_rate": 1.0600000000000002e-05,
"loss": 1.6181,
"step": 68
},
{
"epoch": 0.57,
"learning_rate": 1.04e-05,
"loss": 1.6185,
"step": 69
},
{
"epoch": 0.58,
"learning_rate": 1.02e-05,
"loss": 1.6336,
"step": 70
},
{
"epoch": 0.59,
"learning_rate": 1e-05,
"loss": 1.6343,
"step": 71
},
{
"epoch": 0.6,
"learning_rate": 9.800000000000001e-06,
"loss": 1.612,
"step": 72
},
{
"epoch": 0.61,
"learning_rate": 9.600000000000001e-06,
"loss": 1.633,
"step": 73
},
{
"epoch": 0.62,
"learning_rate": 9.4e-06,
"loss": 1.5895,
"step": 74
},
{
"epoch": 0.62,
"learning_rate": 9.200000000000002e-06,
"loss": 1.6387,
"step": 75
},
{
"epoch": 0.63,
"learning_rate": 9e-06,
"loss": 1.6415,
"step": 76
},
{
"epoch": 0.64,
"learning_rate": 8.8e-06,
"loss": 1.6191,
"step": 77
},
{
"epoch": 0.65,
"learning_rate": 8.6e-06,
"loss": 1.6317,
"step": 78
},
{
"epoch": 0.66,
"learning_rate": 8.400000000000001e-06,
"loss": 1.6239,
"step": 79
},
{
"epoch": 0.67,
"learning_rate": 8.2e-06,
"loss": 1.6315,
"step": 80
},
{
"epoch": 0.67,
"learning_rate": 8.000000000000001e-06,
"loss": 1.5979,
"step": 81
},
{
"epoch": 0.68,
"learning_rate": 7.800000000000002e-06,
"loss": 1.6171,
"step": 82
},
{
"epoch": 0.69,
"learning_rate": 7.600000000000001e-06,
"loss": 1.6199,
"step": 83
},
{
"epoch": 0.7,
"learning_rate": 7.4e-06,
"loss": 1.643,
"step": 84
},
{
"epoch": 0.71,
"learning_rate": 7.2000000000000005e-06,
"loss": 1.6,
"step": 85
},
{
"epoch": 0.72,
"learning_rate": 7e-06,
"loss": 1.6071,
"step": 86
},
{
"epoch": 0.72,
"learning_rate": 6.800000000000001e-06,
"loss": 1.6227,
"step": 87
},
{
"epoch": 0.73,
"learning_rate": 6.600000000000001e-06,
"loss": 1.5972,
"step": 88
},
{
"epoch": 0.74,
"learning_rate": 6.4000000000000006e-06,
"loss": 1.6356,
"step": 89
},
{
"epoch": 0.75,
"learning_rate": 6.200000000000001e-06,
"loss": 1.6205,
"step": 90
},
{
"epoch": 0.76,
"learning_rate": 6e-06,
"loss": 1.5796,
"step": 91
},
{
"epoch": 0.77,
"learning_rate": 5.8e-06,
"loss": 1.6274,
"step": 92
},
{
"epoch": 0.77,
"learning_rate": 5.600000000000001e-06,
"loss": 1.6383,
"step": 93
},
{
"epoch": 0.78,
"learning_rate": 5.400000000000001e-06,
"loss": 1.5923,
"step": 94
},
{
"epoch": 0.79,
"learning_rate": 5.2e-06,
"loss": 1.627,
"step": 95
},
{
"epoch": 0.8,
"learning_rate": 5e-06,
"loss": 1.6201,
"step": 96
},
{
"epoch": 0.81,
"learning_rate": 4.800000000000001e-06,
"loss": 1.6116,
"step": 97
},
{
"epoch": 0.81,
"learning_rate": 4.600000000000001e-06,
"loss": 1.6333,
"step": 98
},
{
"epoch": 0.82,
"learning_rate": 4.4e-06,
"loss": 1.6184,
"step": 99
},
{
"epoch": 0.83,
"learning_rate": 4.2000000000000004e-06,
"loss": 1.5916,
"step": 100
},
{
"epoch": 0.84,
"learning_rate": 4.000000000000001e-06,
"loss": 1.6153,
"step": 101
},
{
"epoch": 0.85,
"learning_rate": 3.8000000000000005e-06,
"loss": 1.5926,
"step": 102
},
{
"epoch": 0.86,
"learning_rate": 3.6000000000000003e-06,
"loss": 1.6308,
"step": 103
},
{
"epoch": 0.86,
"learning_rate": 3.4000000000000005e-06,
"loss": 1.6603,
"step": 104
},
{
"epoch": 0.87,
"learning_rate": 3.2000000000000003e-06,
"loss": 1.6227,
"step": 105
},
{
"epoch": 0.88,
"learning_rate": 3e-06,
"loss": 1.6172,
"step": 106
},
{
"epoch": 0.89,
"learning_rate": 2.8000000000000003e-06,
"loss": 1.6213,
"step": 107
},
{
"epoch": 0.9,
"learning_rate": 2.6e-06,
"loss": 1.6276,
"step": 108
},
{
"epoch": 0.91,
"learning_rate": 2.4000000000000003e-06,
"loss": 1.5951,
"step": 109
},
{
"epoch": 0.91,
"learning_rate": 2.2e-06,
"loss": 1.6211,
"step": 110
},
{
"epoch": 0.92,
"learning_rate": 2.0000000000000003e-06,
"loss": 1.6203,
"step": 111
},
{
"epoch": 0.93,
"learning_rate": 1.8000000000000001e-06,
"loss": 1.6136,
"step": 112
},
{
"epoch": 0.94,
"learning_rate": 1.6000000000000001e-06,
"loss": 1.5951,
"step": 113
},
{
"epoch": 0.95,
"learning_rate": 1.4000000000000001e-06,
"loss": 1.6153,
"step": 114
},
{
"epoch": 0.96,
"learning_rate": 1.2000000000000002e-06,
"loss": 1.6057,
"step": 115
},
{
"epoch": 0.96,
"learning_rate": 1.0000000000000002e-06,
"loss": 1.608,
"step": 116
},
{
"epoch": 0.97,
"learning_rate": 8.000000000000001e-07,
"loss": 1.6029,
"step": 117
},
{
"epoch": 0.98,
"learning_rate": 6.000000000000001e-07,
"loss": 1.6206,
"step": 118
},
{
"epoch": 0.99,
"learning_rate": 4.0000000000000003e-07,
"loss": 1.6221,
"step": 119
},
{
"epoch": 1.0,
"learning_rate": 2.0000000000000002e-07,
"loss": 1.6047,
"step": 120
},
{
"epoch": 1.0,
"step": 120,
"total_flos": 3215233992622080.0,
"train_loss": 1.783738382657369,
"train_runtime": 173137.391,
"train_samples_per_second": 0.178,
"train_steps_per_second": 0.001
}
],
"logging_steps": 1.0,
"max_steps": 120,
"num_train_epochs": 1,
"save_steps": 500,
"total_flos": 3215233992622080.0,
"trial_name": null,
"trial_params": null
}