7067 lines
148 KiB
JSON
7067 lines
148 KiB
JSON
|
|
{
|
||
|
|
"best_global_step": null,
|
||
|
|
"best_metric": null,
|
||
|
|
"best_model_checkpoint": null,
|
||
|
|
"epoch": 1.0,
|
||
|
|
"eval_steps": 250,
|
||
|
|
"global_step": 1000,
|
||
|
|
"is_hyper_param_search": false,
|
||
|
|
"is_local_process_zero": true,
|
||
|
|
"is_world_process_zero": true,
|
||
|
|
"log_history": [
|
||
|
|
{
|
||
|
|
"epoch": 0.001,
|
||
|
|
"grad_norm": 55.0,
|
||
|
|
"learning_rate": 0.0,
|
||
|
|
"loss": 15.8026,
|
||
|
|
"step": 1
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.002,
|
||
|
|
"grad_norm": 53.5,
|
||
|
|
"learning_rate": 2.0000000000000003e-06,
|
||
|
|
"loss": 15.7515,
|
||
|
|
"step": 2
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.003,
|
||
|
|
"grad_norm": 54.0,
|
||
|
|
"learning_rate": 4.000000000000001e-06,
|
||
|
|
"loss": 15.6805,
|
||
|
|
"step": 3
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.004,
|
||
|
|
"grad_norm": 53.75,
|
||
|
|
"learning_rate": 6e-06,
|
||
|
|
"loss": 15.6899,
|
||
|
|
"step": 4
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.005,
|
||
|
|
"grad_norm": 51.75,
|
||
|
|
"learning_rate": 8.000000000000001e-06,
|
||
|
|
"loss": 15.4635,
|
||
|
|
"step": 5
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.006,
|
||
|
|
"grad_norm": 52.0,
|
||
|
|
"learning_rate": 1e-05,
|
||
|
|
"loss": 15.2861,
|
||
|
|
"step": 6
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.007,
|
||
|
|
"grad_norm": 50.25,
|
||
|
|
"learning_rate": 1.2e-05,
|
||
|
|
"loss": 15.0195,
|
||
|
|
"step": 7
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.008,
|
||
|
|
"grad_norm": 51.75,
|
||
|
|
"learning_rate": 1.4000000000000001e-05,
|
||
|
|
"loss": 14.5896,
|
||
|
|
"step": 8
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.009,
|
||
|
|
"grad_norm": 48.5,
|
||
|
|
"learning_rate": 1.6000000000000003e-05,
|
||
|
|
"loss": 14.428,
|
||
|
|
"step": 9
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.01,
|
||
|
|
"grad_norm": 43.0,
|
||
|
|
"learning_rate": 1.8e-05,
|
||
|
|
"loss": 13.8742,
|
||
|
|
"step": 10
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.011,
|
||
|
|
"grad_norm": 35.75,
|
||
|
|
"learning_rate": 2e-05,
|
||
|
|
"loss": 13.1404,
|
||
|
|
"step": 11
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.012,
|
||
|
|
"grad_norm": 25.75,
|
||
|
|
"learning_rate": 2.2000000000000003e-05,
|
||
|
|
"loss": 12.7393,
|
||
|
|
"step": 12
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.013,
|
||
|
|
"grad_norm": 28.25,
|
||
|
|
"learning_rate": 2.4e-05,
|
||
|
|
"loss": 12.4467,
|
||
|
|
"step": 13
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.014,
|
||
|
|
"grad_norm": 39.75,
|
||
|
|
"learning_rate": 2.6000000000000002e-05,
|
||
|
|
"loss": 12.3137,
|
||
|
|
"step": 14
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.015,
|
||
|
|
"grad_norm": 40.75,
|
||
|
|
"learning_rate": 2.8000000000000003e-05,
|
||
|
|
"loss": 11.8957,
|
||
|
|
"step": 15
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.016,
|
||
|
|
"grad_norm": 39.75,
|
||
|
|
"learning_rate": 3e-05,
|
||
|
|
"loss": 11.3899,
|
||
|
|
"step": 16
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.017,
|
||
|
|
"grad_norm": 32.0,
|
||
|
|
"learning_rate": 3.2000000000000005e-05,
|
||
|
|
"loss": 10.4023,
|
||
|
|
"step": 17
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.018,
|
||
|
|
"grad_norm": 25.0,
|
||
|
|
"learning_rate": 3.4000000000000007e-05,
|
||
|
|
"loss": 9.4691,
|
||
|
|
"step": 18
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.019,
|
||
|
|
"grad_norm": 23.5,
|
||
|
|
"learning_rate": 3.6e-05,
|
||
|
|
"loss": 8.8468,
|
||
|
|
"step": 19
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.02,
|
||
|
|
"grad_norm": 20.5,
|
||
|
|
"learning_rate": 3.8e-05,
|
||
|
|
"loss": 8.201,
|
||
|
|
"step": 20
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.021,
|
||
|
|
"grad_norm": 17.5,
|
||
|
|
"learning_rate": 4e-05,
|
||
|
|
"loss": 7.7326,
|
||
|
|
"step": 21
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.022,
|
||
|
|
"grad_norm": 17.375,
|
||
|
|
"learning_rate": 4.2e-05,
|
||
|
|
"loss": 7.3021,
|
||
|
|
"step": 22
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.023,
|
||
|
|
"grad_norm": 14.8125,
|
||
|
|
"learning_rate": 4.4000000000000006e-05,
|
||
|
|
"loss": 7.0761,
|
||
|
|
"step": 23
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.024,
|
||
|
|
"grad_norm": 12.0625,
|
||
|
|
"learning_rate": 4.600000000000001e-05,
|
||
|
|
"loss": 7.0736,
|
||
|
|
"step": 24
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.025,
|
||
|
|
"grad_norm": 9.625,
|
||
|
|
"learning_rate": 4.8e-05,
|
||
|
|
"loss": 6.6617,
|
||
|
|
"step": 25
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.026,
|
||
|
|
"grad_norm": 7.125,
|
||
|
|
"learning_rate": 5e-05,
|
||
|
|
"loss": 6.2627,
|
||
|
|
"step": 26
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.027,
|
||
|
|
"grad_norm": 6.0,
|
||
|
|
"learning_rate": 5.2000000000000004e-05,
|
||
|
|
"loss": 6.108,
|
||
|
|
"step": 27
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.028,
|
||
|
|
"grad_norm": 4.65625,
|
||
|
|
"learning_rate": 5.4000000000000005e-05,
|
||
|
|
"loss": 5.7079,
|
||
|
|
"step": 28
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.029,
|
||
|
|
"grad_norm": 4.5625,
|
||
|
|
"learning_rate": 5.6000000000000006e-05,
|
||
|
|
"loss": 6.0111,
|
||
|
|
"step": 29
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.03,
|
||
|
|
"grad_norm": 3.78125,
|
||
|
|
"learning_rate": 5.8e-05,
|
||
|
|
"loss": 6.0342,
|
||
|
|
"step": 30
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.031,
|
||
|
|
"grad_norm": 3.4375,
|
||
|
|
"learning_rate": 6e-05,
|
||
|
|
"loss": 5.8238,
|
||
|
|
"step": 31
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.032,
|
||
|
|
"grad_norm": 3.4375,
|
||
|
|
"learning_rate": 6.2e-05,
|
||
|
|
"loss": 5.732,
|
||
|
|
"step": 32
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.033,
|
||
|
|
"grad_norm": 3.171875,
|
||
|
|
"learning_rate": 6.400000000000001e-05,
|
||
|
|
"loss": 5.3643,
|
||
|
|
"step": 33
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.034,
|
||
|
|
"grad_norm": 3.546875,
|
||
|
|
"learning_rate": 6.6e-05,
|
||
|
|
"loss": 5.2917,
|
||
|
|
"step": 34
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.035,
|
||
|
|
"grad_norm": 4.21875,
|
||
|
|
"learning_rate": 6.800000000000001e-05,
|
||
|
|
"loss": 5.2339,
|
||
|
|
"step": 35
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.036,
|
||
|
|
"grad_norm": 2.859375,
|
||
|
|
"learning_rate": 7e-05,
|
||
|
|
"loss": 5.2663,
|
||
|
|
"step": 36
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.037,
|
||
|
|
"grad_norm": 2.125,
|
||
|
|
"learning_rate": 7.2e-05,
|
||
|
|
"loss": 5.0012,
|
||
|
|
"step": 37
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.038,
|
||
|
|
"grad_norm": 2.84375,
|
||
|
|
"learning_rate": 7.4e-05,
|
||
|
|
"loss": 4.8893,
|
||
|
|
"step": 38
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.039,
|
||
|
|
"grad_norm": 4.09375,
|
||
|
|
"learning_rate": 7.6e-05,
|
||
|
|
"loss": 4.9856,
|
||
|
|
"step": 39
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.04,
|
||
|
|
"grad_norm": 2.046875,
|
||
|
|
"learning_rate": 7.800000000000001e-05,
|
||
|
|
"loss": 5.0183,
|
||
|
|
"step": 40
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.041,
|
||
|
|
"grad_norm": 8.875,
|
||
|
|
"learning_rate": 8e-05,
|
||
|
|
"loss": 4.9491,
|
||
|
|
"step": 41
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.042,
|
||
|
|
"grad_norm": 4.65625,
|
||
|
|
"learning_rate": 8.2e-05,
|
||
|
|
"loss": 4.7829,
|
||
|
|
"step": 42
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.043,
|
||
|
|
"grad_norm": 4.40625,
|
||
|
|
"learning_rate": 8.4e-05,
|
||
|
|
"loss": 4.9717,
|
||
|
|
"step": 43
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.044,
|
||
|
|
"grad_norm": 4.34375,
|
||
|
|
"learning_rate": 8.6e-05,
|
||
|
|
"loss": 4.817,
|
||
|
|
"step": 44
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.045,
|
||
|
|
"grad_norm": 2.1875,
|
||
|
|
"learning_rate": 8.800000000000001e-05,
|
||
|
|
"loss": 4.8171,
|
||
|
|
"step": 45
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.046,
|
||
|
|
"grad_norm": 5.1875,
|
||
|
|
"learning_rate": 9e-05,
|
||
|
|
"loss": 4.8199,
|
||
|
|
"step": 46
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.047,
|
||
|
|
"grad_norm": 3.40625,
|
||
|
|
"learning_rate": 9.200000000000001e-05,
|
||
|
|
"loss": 4.7965,
|
||
|
|
"step": 47
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.048,
|
||
|
|
"grad_norm": 3.609375,
|
||
|
|
"learning_rate": 9.4e-05,
|
||
|
|
"loss": 4.649,
|
||
|
|
"step": 48
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.049,
|
||
|
|
"grad_norm": 3.59375,
|
||
|
|
"learning_rate": 9.6e-05,
|
||
|
|
"loss": 4.6714,
|
||
|
|
"step": 49
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.05,
|
||
|
|
"grad_norm": 3.25,
|
||
|
|
"learning_rate": 9.8e-05,
|
||
|
|
"loss": 4.5177,
|
||
|
|
"step": 50
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.051,
|
||
|
|
"grad_norm": 3.03125,
|
||
|
|
"learning_rate": 0.0001,
|
||
|
|
"loss": 4.5112,
|
||
|
|
"step": 51
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.052,
|
||
|
|
"grad_norm": 2.15625,
|
||
|
|
"learning_rate": 0.00010200000000000001,
|
||
|
|
"loss": 4.5739,
|
||
|
|
"step": 52
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.053,
|
||
|
|
"grad_norm": 2.1875,
|
||
|
|
"learning_rate": 0.00010400000000000001,
|
||
|
|
"loss": 4.4601,
|
||
|
|
"step": 53
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.054,
|
||
|
|
"grad_norm": 2.46875,
|
||
|
|
"learning_rate": 0.00010600000000000002,
|
||
|
|
"loss": 4.5269,
|
||
|
|
"step": 54
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.055,
|
||
|
|
"grad_norm": 1.8828125,
|
||
|
|
"learning_rate": 0.00010800000000000001,
|
||
|
|
"loss": 4.2753,
|
||
|
|
"step": 55
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.056,
|
||
|
|
"grad_norm": 3.203125,
|
||
|
|
"learning_rate": 0.00011000000000000002,
|
||
|
|
"loss": 4.3292,
|
||
|
|
"step": 56
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.057,
|
||
|
|
"grad_norm": 2.390625,
|
||
|
|
"learning_rate": 0.00011200000000000001,
|
||
|
|
"loss": 4.3489,
|
||
|
|
"step": 57
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.058,
|
||
|
|
"grad_norm": 4.125,
|
||
|
|
"learning_rate": 0.00011399999999999999,
|
||
|
|
"loss": 4.4857,
|
||
|
|
"step": 58
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.059,
|
||
|
|
"grad_norm": 2.96875,
|
||
|
|
"learning_rate": 0.000116,
|
||
|
|
"loss": 4.3047,
|
||
|
|
"step": 59
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.06,
|
||
|
|
"grad_norm": 3.265625,
|
||
|
|
"learning_rate": 0.000118,
|
||
|
|
"loss": 4.4213,
|
||
|
|
"step": 60
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.061,
|
||
|
|
"grad_norm": 11.4375,
|
||
|
|
"learning_rate": 0.00012,
|
||
|
|
"loss": 4.3308,
|
||
|
|
"step": 61
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.062,
|
||
|
|
"grad_norm": 2.59375,
|
||
|
|
"learning_rate": 0.000122,
|
||
|
|
"loss": 4.3524,
|
||
|
|
"step": 62
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.063,
|
||
|
|
"grad_norm": 3.375,
|
||
|
|
"learning_rate": 0.000124,
|
||
|
|
"loss": 4.2236,
|
||
|
|
"step": 63
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.064,
|
||
|
|
"grad_norm": 2.234375,
|
||
|
|
"learning_rate": 0.000126,
|
||
|
|
"loss": 4.3674,
|
||
|
|
"step": 64
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.065,
|
||
|
|
"grad_norm": 3.0,
|
||
|
|
"learning_rate": 0.00012800000000000002,
|
||
|
|
"loss": 4.1765,
|
||
|
|
"step": 65
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.066,
|
||
|
|
"grad_norm": 2.234375,
|
||
|
|
"learning_rate": 0.00013000000000000002,
|
||
|
|
"loss": 4.1043,
|
||
|
|
"step": 66
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.067,
|
||
|
|
"grad_norm": 3.8125,
|
||
|
|
"learning_rate": 0.000132,
|
||
|
|
"loss": 4.0828,
|
||
|
|
"step": 67
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.068,
|
||
|
|
"grad_norm": 2.75,
|
||
|
|
"learning_rate": 0.000134,
|
||
|
|
"loss": 4.1654,
|
||
|
|
"step": 68
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.069,
|
||
|
|
"grad_norm": 3.109375,
|
||
|
|
"learning_rate": 0.00013600000000000003,
|
||
|
|
"loss": 4.0984,
|
||
|
|
"step": 69
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.07,
|
||
|
|
"grad_norm": 2.171875,
|
||
|
|
"learning_rate": 0.000138,
|
||
|
|
"loss": 4.0978,
|
||
|
|
"step": 70
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.071,
|
||
|
|
"grad_norm": 3.84375,
|
||
|
|
"learning_rate": 0.00014,
|
||
|
|
"loss": 4.1044,
|
||
|
|
"step": 71
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.072,
|
||
|
|
"grad_norm": 2.5625,
|
||
|
|
"learning_rate": 0.000142,
|
||
|
|
"loss": 3.967,
|
||
|
|
"step": 72
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.073,
|
||
|
|
"grad_norm": 2.0625,
|
||
|
|
"learning_rate": 0.000144,
|
||
|
|
"loss": 4.0976,
|
||
|
|
"step": 73
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.074,
|
||
|
|
"grad_norm": 2.328125,
|
||
|
|
"learning_rate": 0.000146,
|
||
|
|
"loss": 3.8734,
|
||
|
|
"step": 74
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.075,
|
||
|
|
"grad_norm": 2.015625,
|
||
|
|
"learning_rate": 0.000148,
|
||
|
|
"loss": 3.9381,
|
||
|
|
"step": 75
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.076,
|
||
|
|
"grad_norm": 2.921875,
|
||
|
|
"learning_rate": 0.00015000000000000001,
|
||
|
|
"loss": 3.97,
|
||
|
|
"step": 76
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.077,
|
||
|
|
"grad_norm": 2.046875,
|
||
|
|
"learning_rate": 0.000152,
|
||
|
|
"loss": 3.9806,
|
||
|
|
"step": 77
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.078,
|
||
|
|
"grad_norm": 3.25,
|
||
|
|
"learning_rate": 0.000154,
|
||
|
|
"loss": 3.7804,
|
||
|
|
"step": 78
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.079,
|
||
|
|
"grad_norm": 2.578125,
|
||
|
|
"learning_rate": 0.00015600000000000002,
|
||
|
|
"loss": 3.9961,
|
||
|
|
"step": 79
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.08,
|
||
|
|
"grad_norm": 1.9765625,
|
||
|
|
"learning_rate": 0.00015800000000000002,
|
||
|
|
"loss": 3.9899,
|
||
|
|
"step": 80
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.081,
|
||
|
|
"grad_norm": 2.171875,
|
||
|
|
"learning_rate": 0.00016,
|
||
|
|
"loss": 3.8394,
|
||
|
|
"step": 81
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.082,
|
||
|
|
"grad_norm": 1.8671875,
|
||
|
|
"learning_rate": 0.000162,
|
||
|
|
"loss": 3.8614,
|
||
|
|
"step": 82
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.083,
|
||
|
|
"grad_norm": 2.140625,
|
||
|
|
"learning_rate": 0.000164,
|
||
|
|
"loss": 3.7687,
|
||
|
|
"step": 83
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.084,
|
||
|
|
"grad_norm": 1.5546875,
|
||
|
|
"learning_rate": 0.000166,
|
||
|
|
"loss": 3.8413,
|
||
|
|
"step": 84
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.085,
|
||
|
|
"grad_norm": 2.375,
|
||
|
|
"learning_rate": 0.000168,
|
||
|
|
"loss": 3.7533,
|
||
|
|
"step": 85
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.086,
|
||
|
|
"grad_norm": 1.890625,
|
||
|
|
"learning_rate": 0.00017,
|
||
|
|
"loss": 3.8058,
|
||
|
|
"step": 86
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.087,
|
||
|
|
"grad_norm": 1.90625,
|
||
|
|
"learning_rate": 0.000172,
|
||
|
|
"loss": 3.8243,
|
||
|
|
"step": 87
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.088,
|
||
|
|
"grad_norm": 2.25,
|
||
|
|
"learning_rate": 0.000174,
|
||
|
|
"loss": 3.7795,
|
||
|
|
"step": 88
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.089,
|
||
|
|
"grad_norm": 1.8359375,
|
||
|
|
"learning_rate": 0.00017600000000000002,
|
||
|
|
"loss": 3.7421,
|
||
|
|
"step": 89
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.09,
|
||
|
|
"grad_norm": 1.7421875,
|
||
|
|
"learning_rate": 0.00017800000000000002,
|
||
|
|
"loss": 3.7222,
|
||
|
|
"step": 90
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.091,
|
||
|
|
"grad_norm": 1.5078125,
|
||
|
|
"learning_rate": 0.00018,
|
||
|
|
"loss": 3.727,
|
||
|
|
"step": 91
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.092,
|
||
|
|
"grad_norm": 2.328125,
|
||
|
|
"learning_rate": 0.000182,
|
||
|
|
"loss": 3.8749,
|
||
|
|
"step": 92
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.093,
|
||
|
|
"grad_norm": 1.7734375,
|
||
|
|
"learning_rate": 0.00018400000000000003,
|
||
|
|
"loss": 3.6911,
|
||
|
|
"step": 93
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.094,
|
||
|
|
"grad_norm": 2.4375,
|
||
|
|
"learning_rate": 0.00018600000000000002,
|
||
|
|
"loss": 3.6221,
|
||
|
|
"step": 94
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.095,
|
||
|
|
"grad_norm": 2.265625,
|
||
|
|
"learning_rate": 0.000188,
|
||
|
|
"loss": 3.6225,
|
||
|
|
"step": 95
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.096,
|
||
|
|
"grad_norm": 2.078125,
|
||
|
|
"learning_rate": 0.00019,
|
||
|
|
"loss": 3.5456,
|
||
|
|
"step": 96
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.097,
|
||
|
|
"grad_norm": 3.0625,
|
||
|
|
"learning_rate": 0.000192,
|
||
|
|
"loss": 3.5316,
|
||
|
|
"step": 97
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.098,
|
||
|
|
"grad_norm": 1.75,
|
||
|
|
"learning_rate": 0.000194,
|
||
|
|
"loss": 3.5942,
|
||
|
|
"step": 98
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.099,
|
||
|
|
"grad_norm": 1.5078125,
|
||
|
|
"learning_rate": 0.000196,
|
||
|
|
"loss": 3.5513,
|
||
|
|
"step": 99
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.1,
|
||
|
|
"grad_norm": 2.8125,
|
||
|
|
"learning_rate": 0.00019800000000000002,
|
||
|
|
"loss": 3.673,
|
||
|
|
"step": 100
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.101,
|
||
|
|
"grad_norm": 2.328125,
|
||
|
|
"learning_rate": 0.0002,
|
||
|
|
"loss": 3.6088,
|
||
|
|
"step": 101
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.102,
|
||
|
|
"grad_norm": 1.4921875,
|
||
|
|
"learning_rate": 0.00019999939076577905,
|
||
|
|
"loss": 3.5971,
|
||
|
|
"step": 102
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.103,
|
||
|
|
"grad_norm": 1.421875,
|
||
|
|
"learning_rate": 0.00019999756307053948,
|
||
|
|
"loss": 3.5318,
|
||
|
|
"step": 103
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.104,
|
||
|
|
"grad_norm": 2.046875,
|
||
|
|
"learning_rate": 0.00019999451693655123,
|
||
|
|
"loss": 3.5199,
|
||
|
|
"step": 104
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.105,
|
||
|
|
"grad_norm": 1.7734375,
|
||
|
|
"learning_rate": 0.00019999025240093044,
|
||
|
|
"loss": 3.5049,
|
||
|
|
"step": 105
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.106,
|
||
|
|
"grad_norm": 2.234375,
|
||
|
|
"learning_rate": 0.00019998476951563915,
|
||
|
|
"loss": 3.5481,
|
||
|
|
"step": 106
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.107,
|
||
|
|
"grad_norm": 1.9921875,
|
||
|
|
"learning_rate": 0.00019997806834748456,
|
||
|
|
"loss": 3.4391,
|
||
|
|
"step": 107
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.108,
|
||
|
|
"grad_norm": 1.4609375,
|
||
|
|
"learning_rate": 0.00019997014897811833,
|
||
|
|
"loss": 3.5096,
|
||
|
|
"step": 108
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.109,
|
||
|
|
"grad_norm": 1.8046875,
|
||
|
|
"learning_rate": 0.00019996101150403543,
|
||
|
|
"loss": 3.424,
|
||
|
|
"step": 109
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.11,
|
||
|
|
"grad_norm": 2.4375,
|
||
|
|
"learning_rate": 0.00019995065603657316,
|
||
|
|
"loss": 3.4862,
|
||
|
|
"step": 110
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.111,
|
||
|
|
"grad_norm": 1.5859375,
|
||
|
|
"learning_rate": 0.0001999390827019096,
|
||
|
|
"loss": 3.5117,
|
||
|
|
"step": 111
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.112,
|
||
|
|
"grad_norm": 1.7265625,
|
||
|
|
"learning_rate": 0.0001999262916410621,
|
||
|
|
"loss": 3.452,
|
||
|
|
"step": 112
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.113,
|
||
|
|
"grad_norm": 1.7578125,
|
||
|
|
"learning_rate": 0.00019991228300988585,
|
||
|
|
"loss": 3.4834,
|
||
|
|
"step": 113
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.114,
|
||
|
|
"grad_norm": 1.8984375,
|
||
|
|
"learning_rate": 0.00019989705697907149,
|
||
|
|
"loss": 3.4673,
|
||
|
|
"step": 114
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.115,
|
||
|
|
"grad_norm": 1.625,
|
||
|
|
"learning_rate": 0.0001998806137341434,
|
||
|
|
"loss": 3.2031,
|
||
|
|
"step": 115
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.116,
|
||
|
|
"grad_norm": 1.6015625,
|
||
|
|
"learning_rate": 0.0001998629534754574,
|
||
|
|
"loss": 3.3329,
|
||
|
|
"step": 116
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.117,
|
||
|
|
"grad_norm": 1.4921875,
|
||
|
|
"learning_rate": 0.00019984407641819812,
|
||
|
|
"loss": 3.4215,
|
||
|
|
"step": 117
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.118,
|
||
|
|
"grad_norm": 1.65625,
|
||
|
|
"learning_rate": 0.00019982398279237655,
|
||
|
|
"loss": 3.3886,
|
||
|
|
"step": 118
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.119,
|
||
|
|
"grad_norm": 1.421875,
|
||
|
|
"learning_rate": 0.00019980267284282717,
|
||
|
|
"loss": 3.2904,
|
||
|
|
"step": 119
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.12,
|
||
|
|
"grad_norm": 1.3515625,
|
||
|
|
"learning_rate": 0.000199780146829205,
|
||
|
|
"loss": 3.3135,
|
||
|
|
"step": 120
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.121,
|
||
|
|
"grad_norm": 1.8671875,
|
||
|
|
"learning_rate": 0.00019975640502598244,
|
||
|
|
"loss": 3.4523,
|
||
|
|
"step": 121
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.122,
|
||
|
|
"grad_norm": 1.796875,
|
||
|
|
"learning_rate": 0.00019973144772244582,
|
||
|
|
"loss": 3.248,
|
||
|
|
"step": 122
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.123,
|
||
|
|
"grad_norm": 1.4140625,
|
||
|
|
"learning_rate": 0.00019970527522269205,
|
||
|
|
"loss": 3.3419,
|
||
|
|
"step": 123
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.124,
|
||
|
|
"grad_norm": 1.453125,
|
||
|
|
"learning_rate": 0.00019967788784562473,
|
||
|
|
"loss": 3.296,
|
||
|
|
"step": 124
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.125,
|
||
|
|
"grad_norm": 1.90625,
|
||
|
|
"learning_rate": 0.00019964928592495045,
|
||
|
|
"loss": 3.2151,
|
||
|
|
"step": 125
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.126,
|
||
|
|
"grad_norm": 1.3828125,
|
||
|
|
"learning_rate": 0.00019961946980917456,
|
||
|
|
"loss": 3.3786,
|
||
|
|
"step": 126
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.127,
|
||
|
|
"grad_norm": 1.859375,
|
||
|
|
"learning_rate": 0.00019958843986159704,
|
||
|
|
"loss": 3.4188,
|
||
|
|
"step": 127
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.128,
|
||
|
|
"grad_norm": 1.46875,
|
||
|
|
"learning_rate": 0.00019955619646030802,
|
||
|
|
"loss": 3.4489,
|
||
|
|
"step": 128
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.129,
|
||
|
|
"grad_norm": 1.7265625,
|
||
|
|
"learning_rate": 0.0001995227399981831,
|
||
|
|
"loss": 3.3746,
|
||
|
|
"step": 129
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.13,
|
||
|
|
"grad_norm": 1.921875,
|
||
|
|
"learning_rate": 0.00019948807088287883,
|
||
|
|
"loss": 3.29,
|
||
|
|
"step": 130
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.131,
|
||
|
|
"grad_norm": 1.3984375,
|
||
|
|
"learning_rate": 0.00019945218953682734,
|
||
|
|
"loss": 3.2717,
|
||
|
|
"step": 131
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.132,
|
||
|
|
"grad_norm": 4.21875,
|
||
|
|
"learning_rate": 0.00019941509639723155,
|
||
|
|
"loss": 3.2601,
|
||
|
|
"step": 132
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.133,
|
||
|
|
"grad_norm": 2.15625,
|
||
|
|
"learning_rate": 0.00019937679191605963,
|
||
|
|
"loss": 3.3615,
|
||
|
|
"step": 133
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.134,
|
||
|
|
"grad_norm": 2.390625,
|
||
|
|
"learning_rate": 0.00019933727656003963,
|
||
|
|
"loss": 3.4326,
|
||
|
|
"step": 134
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.135,
|
||
|
|
"grad_norm": 1.9296875,
|
||
|
|
"learning_rate": 0.0001992965508106537,
|
||
|
|
"loss": 3.3733,
|
||
|
|
"step": 135
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.136,
|
||
|
|
"grad_norm": 1.6953125,
|
||
|
|
"learning_rate": 0.00019925461516413223,
|
||
|
|
"loss": 3.3488,
|
||
|
|
"step": 136
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.137,
|
||
|
|
"grad_norm": 1.46875,
|
||
|
|
"learning_rate": 0.0001992114701314478,
|
||
|
|
"loss": 3.2991,
|
||
|
|
"step": 137
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.138,
|
||
|
|
"grad_norm": 1.375,
|
||
|
|
"learning_rate": 0.00019916711623830903,
|
||
|
|
"loss": 3.2037,
|
||
|
|
"step": 138
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.139,
|
||
|
|
"grad_norm": 1.5234375,
|
||
|
|
"learning_rate": 0.00019912155402515417,
|
||
|
|
"loss": 3.2748,
|
||
|
|
"step": 139
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.14,
|
||
|
|
"grad_norm": 1.4375,
|
||
|
|
"learning_rate": 0.00019907478404714436,
|
||
|
|
"loss": 3.2753,
|
||
|
|
"step": 140
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.141,
|
||
|
|
"grad_norm": 2.0,
|
||
|
|
"learning_rate": 0.00019902680687415705,
|
||
|
|
"loss": 3.3406,
|
||
|
|
"step": 141
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.142,
|
||
|
|
"grad_norm": 1.4296875,
|
||
|
|
"learning_rate": 0.0001989776230907789,
|
||
|
|
"loss": 3.3422,
|
||
|
|
"step": 142
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.143,
|
||
|
|
"grad_norm": 1.703125,
|
||
|
|
"learning_rate": 0.00019892723329629887,
|
||
|
|
"loss": 3.4146,
|
||
|
|
"step": 143
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.144,
|
||
|
|
"grad_norm": 1.4453125,
|
||
|
|
"learning_rate": 0.0001988756381047006,
|
||
|
|
"loss": 3.3321,
|
||
|
|
"step": 144
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.145,
|
||
|
|
"grad_norm": 1.46875,
|
||
|
|
"learning_rate": 0.0001988228381446553,
|
||
|
|
"loss": 3.3225,
|
||
|
|
"step": 145
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.146,
|
||
|
|
"grad_norm": 1.3984375,
|
||
|
|
"learning_rate": 0.00019876883405951377,
|
||
|
|
"loss": 3.2079,
|
||
|
|
"step": 146
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.147,
|
||
|
|
"grad_norm": 1.59375,
|
||
|
|
"learning_rate": 0.0001987136265072988,
|
||
|
|
"loss": 3.1757,
|
||
|
|
"step": 147
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.148,
|
||
|
|
"grad_norm": 1.40625,
|
||
|
|
"learning_rate": 0.00019865721616069696,
|
||
|
|
"loss": 3.1882,
|
||
|
|
"step": 148
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.149,
|
||
|
|
"grad_norm": 1.3671875,
|
||
|
|
"learning_rate": 0.0001985996037070505,
|
||
|
|
"loss": 3.2314,
|
||
|
|
"step": 149
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.15,
|
||
|
|
"grad_norm": 1.2109375,
|
||
|
|
"learning_rate": 0.00019854078984834903,
|
||
|
|
"loss": 3.1942,
|
||
|
|
"step": 150
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.151,
|
||
|
|
"grad_norm": 1.4453125,
|
||
|
|
"learning_rate": 0.00019848077530122083,
|
||
|
|
"loss": 3.2477,
|
||
|
|
"step": 151
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.152,
|
||
|
|
"grad_norm": 1.7734375,
|
||
|
|
"learning_rate": 0.0001984195607969242,
|
||
|
|
"loss": 3.1375,
|
||
|
|
"step": 152
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.153,
|
||
|
|
"grad_norm": 1.484375,
|
||
|
|
"learning_rate": 0.00019835714708133862,
|
||
|
|
"loss": 3.145,
|
||
|
|
"step": 153
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.154,
|
||
|
|
"grad_norm": 1.2265625,
|
||
|
|
"learning_rate": 0.00019829353491495545,
|
||
|
|
"loss": 3.1747,
|
||
|
|
"step": 154
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.155,
|
||
|
|
"grad_norm": 1.3046875,
|
||
|
|
"learning_rate": 0.0001982287250728689,
|
||
|
|
"loss": 3.2027,
|
||
|
|
"step": 155
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.156,
|
||
|
|
"grad_norm": 1.4609375,
|
||
|
|
"learning_rate": 0.00019816271834476642,
|
||
|
|
"loss": 3.09,
|
||
|
|
"step": 156
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.157,
|
||
|
|
"grad_norm": 2.578125,
|
||
|
|
"learning_rate": 0.00019809551553491916,
|
||
|
|
"loss": 3.2477,
|
||
|
|
"step": 157
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.158,
|
||
|
|
"grad_norm": 1.7421875,
|
||
|
|
"learning_rate": 0.00019802711746217218,
|
||
|
|
"loss": 3.1619,
|
||
|
|
"step": 158
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.159,
|
||
|
|
"grad_norm": 1.3984375,
|
||
|
|
"learning_rate": 0.0001979575249599344,
|
||
|
|
"loss": 3.2682,
|
||
|
|
"step": 159
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.16,
|
||
|
|
"grad_norm": 1.6328125,
|
||
|
|
"learning_rate": 0.0001978867388761685,
|
||
|
|
"loss": 3.0885,
|
||
|
|
"step": 160
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.161,
|
||
|
|
"grad_norm": 1.2890625,
|
||
|
|
"learning_rate": 0.00019781476007338058,
|
||
|
|
"loss": 3.2491,
|
||
|
|
"step": 161
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.162,
|
||
|
|
"grad_norm": 1.5,
|
||
|
|
"learning_rate": 0.0001977415894286096,
|
||
|
|
"loss": 3.21,
|
||
|
|
"step": 162
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.163,
|
||
|
|
"grad_norm": 1.2734375,
|
||
|
|
"learning_rate": 0.0001976672278334168,
|
||
|
|
"loss": 3.1559,
|
||
|
|
"step": 163
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.164,
|
||
|
|
"grad_norm": 1.3515625,
|
||
|
|
"learning_rate": 0.00019759167619387476,
|
||
|
|
"loss": 3.0837,
|
||
|
|
"step": 164
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.165,
|
||
|
|
"grad_norm": 1.3046875,
|
||
|
|
"learning_rate": 0.00019751493543055632,
|
||
|
|
"loss": 3.1893,
|
||
|
|
"step": 165
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.166,
|
||
|
|
"grad_norm": 1.203125,
|
||
|
|
"learning_rate": 0.00019743700647852354,
|
||
|
|
"loss": 3.1023,
|
||
|
|
"step": 166
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.167,
|
||
|
|
"grad_norm": 1.1953125,
|
||
|
|
"learning_rate": 0.00019735789028731604,
|
||
|
|
"loss": 3.0936,
|
||
|
|
"step": 167
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.168,
|
||
|
|
"grad_norm": 1.359375,
|
||
|
|
"learning_rate": 0.00019727758782093967,
|
||
|
|
"loss": 3.2097,
|
||
|
|
"step": 168
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.169,
|
||
|
|
"grad_norm": 1.6171875,
|
||
|
|
"learning_rate": 0.00019719610005785465,
|
||
|
|
"loss": 3.1198,
|
||
|
|
"step": 169
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.17,
|
||
|
|
"grad_norm": 1.3359375,
|
||
|
|
"learning_rate": 0.00019711342799096361,
|
||
|
|
"loss": 3.1823,
|
||
|
|
"step": 170
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.171,
|
||
|
|
"grad_norm": 1.0703125,
|
||
|
|
"learning_rate": 0.00019702957262759965,
|
||
|
|
"loss": 3.1136,
|
||
|
|
"step": 171
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.172,
|
||
|
|
"grad_norm": 4.90625,
|
||
|
|
"learning_rate": 0.0001969445349895139,
|
||
|
|
"loss": 3.1076,
|
||
|
|
"step": 172
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.173,
|
||
|
|
"grad_norm": 1.625,
|
||
|
|
"learning_rate": 0.0001968583161128631,
|
||
|
|
"loss": 3.113,
|
||
|
|
"step": 173
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.174,
|
||
|
|
"grad_norm": 1.6640625,
|
||
|
|
"learning_rate": 0.00019677091704819715,
|
||
|
|
"loss": 3.1752,
|
||
|
|
"step": 174
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.175,
|
||
|
|
"grad_norm": 1.5390625,
|
||
|
|
"learning_rate": 0.00019668233886044597,
|
||
|
|
"loss": 3.1309,
|
||
|
|
"step": 175
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.176,
|
||
|
|
"grad_norm": 1.34375,
|
||
|
|
"learning_rate": 0.00019659258262890683,
|
||
|
|
"loss": 3.1505,
|
||
|
|
"step": 176
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.177,
|
||
|
|
"grad_norm": 1.3125,
|
||
|
|
"learning_rate": 0.00019650164944723115,
|
||
|
|
"loss": 3.1151,
|
||
|
|
"step": 177
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.178,
|
||
|
|
"grad_norm": 1.5546875,
|
||
|
|
"learning_rate": 0.00019640954042341103,
|
||
|
|
"loss": 3.085,
|
||
|
|
"step": 178
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.179,
|
||
|
|
"grad_norm": 1.421875,
|
||
|
|
"learning_rate": 0.00019631625667976583,
|
||
|
|
"loss": 3.0857,
|
||
|
|
"step": 179
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.18,
|
||
|
|
"grad_norm": 1.1953125,
|
||
|
|
"learning_rate": 0.00019622179935292855,
|
||
|
|
"loss": 3.2549,
|
||
|
|
"step": 180
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.181,
|
||
|
|
"grad_norm": 1.265625,
|
||
|
|
"learning_rate": 0.0001961261695938319,
|
||
|
|
"loss": 3.2138,
|
||
|
|
"step": 181
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.182,
|
||
|
|
"grad_norm": 1.1796875,
|
||
|
|
"learning_rate": 0.0001960293685676943,
|
||
|
|
"loss": 3.0507,
|
||
|
|
"step": 182
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.183,
|
||
|
|
"grad_norm": 1.4453125,
|
||
|
|
"learning_rate": 0.00019593139745400576,
|
||
|
|
"loss": 3.2578,
|
||
|
|
"step": 183
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.184,
|
||
|
|
"grad_norm": 1.2109375,
|
||
|
|
"learning_rate": 0.00019583225744651333,
|
||
|
|
"loss": 3.0951,
|
||
|
|
"step": 184
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.185,
|
||
|
|
"grad_norm": 1.6875,
|
||
|
|
"learning_rate": 0.00019573194975320673,
|
||
|
|
"loss": 3.1273,
|
||
|
|
"step": 185
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.186,
|
||
|
|
"grad_norm": 1.296875,
|
||
|
|
"learning_rate": 0.00019563047559630357,
|
||
|
|
"loss": 3.0946,
|
||
|
|
"step": 186
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.187,
|
||
|
|
"grad_norm": 1.21875,
|
||
|
|
"learning_rate": 0.00019552783621223436,
|
||
|
|
"loss": 3.1003,
|
||
|
|
"step": 187
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.188,
|
||
|
|
"grad_norm": 2.359375,
|
||
|
|
"learning_rate": 0.0001954240328516277,
|
||
|
|
"loss": 3.2026,
|
||
|
|
"step": 188
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.189,
|
||
|
|
"grad_norm": 1.515625,
|
||
|
|
"learning_rate": 0.0001953190667792947,
|
||
|
|
"loss": 3.1465,
|
||
|
|
"step": 189
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.19,
|
||
|
|
"grad_norm": 1.5390625,
|
||
|
|
"learning_rate": 0.00019521293927421388,
|
||
|
|
"loss": 3.1967,
|
||
|
|
"step": 190
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.191,
|
||
|
|
"grad_norm": 1.2109375,
|
||
|
|
"learning_rate": 0.00019510565162951537,
|
||
|
|
"loss": 3.1578,
|
||
|
|
"step": 191
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.192,
|
||
|
|
"grad_norm": 1.296875,
|
||
|
|
"learning_rate": 0.00019499720515246525,
|
||
|
|
"loss": 3.1388,
|
||
|
|
"step": 192
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.193,
|
||
|
|
"grad_norm": 1.140625,
|
||
|
|
"learning_rate": 0.00019488760116444966,
|
||
|
|
"loss": 3.1873,
|
||
|
|
"step": 193
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.194,
|
||
|
|
"grad_norm": 1.5703125,
|
||
|
|
"learning_rate": 0.0001947768410009586,
|
||
|
|
"loss": 3.1005,
|
||
|
|
"step": 194
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.195,
|
||
|
|
"grad_norm": 1.515625,
|
||
|
|
"learning_rate": 0.00019466492601156966,
|
||
|
|
"loss": 3.0942,
|
||
|
|
"step": 195
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.196,
|
||
|
|
"grad_norm": 1.25,
|
||
|
|
"learning_rate": 0.0001945518575599317,
|
||
|
|
"loss": 3.0119,
|
||
|
|
"step": 196
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.197,
|
||
|
|
"grad_norm": 1.1171875,
|
||
|
|
"learning_rate": 0.00019443763702374812,
|
||
|
|
"loss": 3.0152,
|
||
|
|
"step": 197
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.198,
|
||
|
|
"grad_norm": 1.609375,
|
||
|
|
"learning_rate": 0.0001943222657947601,
|
||
|
|
"loss": 2.9887,
|
||
|
|
"step": 198
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.199,
|
||
|
|
"grad_norm": 1.421875,
|
||
|
|
"learning_rate": 0.00019420574527872968,
|
||
|
|
"loss": 3.0333,
|
||
|
|
"step": 199
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.2,
|
||
|
|
"grad_norm": 1.1171875,
|
||
|
|
"learning_rate": 0.00019408807689542257,
|
||
|
|
"loss": 3.1974,
|
||
|
|
"step": 200
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.201,
|
||
|
|
"grad_norm": 1.4453125,
|
||
|
|
"learning_rate": 0.00019396926207859084,
|
||
|
|
"loss": 3.1598,
|
||
|
|
"step": 201
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.202,
|
||
|
|
"grad_norm": 1.21875,
|
||
|
|
"learning_rate": 0.0001938493022759556,
|
||
|
|
"loss": 3.1167,
|
||
|
|
"step": 202
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.203,
|
||
|
|
"grad_norm": 1.171875,
|
||
|
|
"learning_rate": 0.00019372819894918915,
|
||
|
|
"loss": 3.0386,
|
||
|
|
"step": 203
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.204,
|
||
|
|
"grad_norm": 1.890625,
|
||
|
|
"learning_rate": 0.00019360595357389735,
|
||
|
|
"loss": 3.0719,
|
||
|
|
"step": 204
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.205,
|
||
|
|
"grad_norm": 1.328125,
|
||
|
|
"learning_rate": 0.00019348256763960145,
|
||
|
|
"loss": 3.1225,
|
||
|
|
"step": 205
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.206,
|
||
|
|
"grad_norm": 1.578125,
|
||
|
|
"learning_rate": 0.00019335804264972018,
|
||
|
|
"loss": 3.1224,
|
||
|
|
"step": 206
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.207,
|
||
|
|
"grad_norm": 1.4140625,
|
||
|
|
"learning_rate": 0.00019323238012155123,
|
||
|
|
"loss": 3.0044,
|
||
|
|
"step": 207
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.208,
|
||
|
|
"grad_norm": 1.25,
|
||
|
|
"learning_rate": 0.00019310558158625285,
|
||
|
|
"loss": 3.0157,
|
||
|
|
"step": 208
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.209,
|
||
|
|
"grad_norm": 1.375,
|
||
|
|
"learning_rate": 0.00019297764858882514,
|
||
|
|
"loss": 2.9538,
|
||
|
|
"step": 209
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.21,
|
||
|
|
"grad_norm": 1.375,
|
||
|
|
"learning_rate": 0.00019284858268809137,
|
||
|
|
"loss": 3.0122,
|
||
|
|
"step": 210
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.211,
|
||
|
|
"grad_norm": 1.3359375,
|
||
|
|
"learning_rate": 0.00019271838545667876,
|
||
|
|
"loss": 3.0153,
|
||
|
|
"step": 211
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.212,
|
||
|
|
"grad_norm": 1.015625,
|
||
|
|
"learning_rate": 0.0001925870584809995,
|
||
|
|
"loss": 3.0157,
|
||
|
|
"step": 212
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.213,
|
||
|
|
"grad_norm": 1.484375,
|
||
|
|
"learning_rate": 0.00019245460336123134,
|
||
|
|
"loss": 3.1123,
|
||
|
|
"step": 213
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.214,
|
||
|
|
"grad_norm": 1.1171875,
|
||
|
|
"learning_rate": 0.00019232102171129811,
|
||
|
|
"loss": 3.0119,
|
||
|
|
"step": 214
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.215,
|
||
|
|
"grad_norm": 1.109375,
|
||
|
|
"learning_rate": 0.00019218631515885006,
|
||
|
|
"loss": 3.0785,
|
||
|
|
"step": 215
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.216,
|
||
|
|
"grad_norm": 1.15625,
|
||
|
|
"learning_rate": 0.00019205048534524406,
|
||
|
|
"loss": 2.9962,
|
||
|
|
"step": 216
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.217,
|
||
|
|
"grad_norm": 1.046875,
|
||
|
|
"learning_rate": 0.00019191353392552344,
|
||
|
|
"loss": 2.98,
|
||
|
|
"step": 217
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.218,
|
||
|
|
"grad_norm": 1.2890625,
|
||
|
|
"learning_rate": 0.00019177546256839812,
|
||
|
|
"loss": 3.1094,
|
||
|
|
"step": 218
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.219,
|
||
|
|
"grad_norm": 1.84375,
|
||
|
|
"learning_rate": 0.00019163627295622397,
|
||
|
|
"loss": 3.0632,
|
||
|
|
"step": 219
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.22,
|
||
|
|
"grad_norm": 1.3515625,
|
||
|
|
"learning_rate": 0.0001914959667849825,
|
||
|
|
"loss": 3.0286,
|
||
|
|
"step": 220
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.221,
|
||
|
|
"grad_norm": 1.3359375,
|
||
|
|
"learning_rate": 0.0001913545457642601,
|
||
|
|
"loss": 3.2119,
|
||
|
|
"step": 221
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.222,
|
||
|
|
"grad_norm": 1.203125,
|
||
|
|
"learning_rate": 0.0001912120116172273,
|
||
|
|
"loss": 3.1813,
|
||
|
|
"step": 222
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.223,
|
||
|
|
"grad_norm": 1.1640625,
|
||
|
|
"learning_rate": 0.00019106836608061772,
|
||
|
|
"loss": 3.144,
|
||
|
|
"step": 223
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.224,
|
||
|
|
"grad_norm": 1.2421875,
|
||
|
|
"learning_rate": 0.00019092361090470688,
|
||
|
|
"loss": 3.012,
|
||
|
|
"step": 224
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.225,
|
||
|
|
"grad_norm": 1.1328125,
|
||
|
|
"learning_rate": 0.00019077774785329087,
|
||
|
|
"loss": 3.0367,
|
||
|
|
"step": 225
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.226,
|
||
|
|
"grad_norm": 1.0390625,
|
||
|
|
"learning_rate": 0.000190630778703665,
|
||
|
|
"loss": 3.0175,
|
||
|
|
"step": 226
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.227,
|
||
|
|
"grad_norm": 1.296875,
|
||
|
|
"learning_rate": 0.00019048270524660196,
|
||
|
|
"loss": 2.9683,
|
||
|
|
"step": 227
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.228,
|
||
|
|
"grad_norm": 1.140625,
|
||
|
|
"learning_rate": 0.0001903335292863301,
|
||
|
|
"loss": 3.0164,
|
||
|
|
"step": 228
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.229,
|
||
|
|
"grad_norm": 1.234375,
|
||
|
|
"learning_rate": 0.0001901832526405114,
|
||
|
|
"loss": 2.9548,
|
||
|
|
"step": 229
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.23,
|
||
|
|
"grad_norm": 1.6875,
|
||
|
|
"learning_rate": 0.00019003187714021938,
|
||
|
|
"loss": 2.4954,
|
||
|
|
"step": 230
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.231,
|
||
|
|
"grad_norm": 5.59375,
|
||
|
|
"learning_rate": 0.0001898794046299167,
|
||
|
|
"loss": 2.6911,
|
||
|
|
"step": 231
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.232,
|
||
|
|
"grad_norm": 132.0,
|
||
|
|
"learning_rate": 0.00018972583696743285,
|
||
|
|
"loss": 3.6355,
|
||
|
|
"step": 232
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.233,
|
||
|
|
"grad_norm": 2.875,
|
||
|
|
"learning_rate": 0.0001895711760239413,
|
||
|
|
"loss": 3.0587,
|
||
|
|
"step": 233
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.234,
|
||
|
|
"grad_norm": 1.7734375,
|
||
|
|
"learning_rate": 0.0001894154236839368,
|
||
|
|
"loss": 3.0121,
|
||
|
|
"step": 234
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.235,
|
||
|
|
"grad_norm": 52.0,
|
||
|
|
"learning_rate": 0.00018925858184521256,
|
||
|
|
"loss": 3.0583,
|
||
|
|
"step": 235
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.236,
|
||
|
|
"grad_norm": 2.484375,
|
||
|
|
"learning_rate": 0.0001891006524188368,
|
||
|
|
"loss": 3.1517,
|
||
|
|
"step": 236
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.237,
|
||
|
|
"grad_norm": 1.7578125,
|
||
|
|
"learning_rate": 0.00018894163732912977,
|
||
|
|
"loss": 3.1375,
|
||
|
|
"step": 237
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.238,
|
||
|
|
"grad_norm": 1.59375,
|
||
|
|
"learning_rate": 0.00018878153851364013,
|
||
|
|
"loss": 3.1155,
|
||
|
|
"step": 238
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.239,
|
||
|
|
"grad_norm": 1.546875,
|
||
|
|
"learning_rate": 0.00018862035792312147,
|
||
|
|
"loss": 3.0236,
|
||
|
|
"step": 239
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.24,
|
||
|
|
"grad_norm": 1.328125,
|
||
|
|
"learning_rate": 0.0001884580975215084,
|
||
|
|
"loss": 3.2067,
|
||
|
|
"step": 240
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.241,
|
||
|
|
"grad_norm": 1.2109375,
|
||
|
|
"learning_rate": 0.00018829475928589271,
|
||
|
|
"loss": 3.1324,
|
||
|
|
"step": 241
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.242,
|
||
|
|
"grad_norm": 1.1875,
|
||
|
|
"learning_rate": 0.0001881303452064992,
|
||
|
|
"loss": 3.0433,
|
||
|
|
"step": 242
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.243,
|
||
|
|
"grad_norm": 1.1640625,
|
||
|
|
"learning_rate": 0.00018796485728666165,
|
||
|
|
"loss": 2.9619,
|
||
|
|
"step": 243
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.244,
|
||
|
|
"grad_norm": 1.09375,
|
||
|
|
"learning_rate": 0.00018779829754279805,
|
||
|
|
"loss": 3.052,
|
||
|
|
"step": 244
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.245,
|
||
|
|
"grad_norm": 1.2109375,
|
||
|
|
"learning_rate": 0.00018763066800438636,
|
||
|
|
"loss": 3.0409,
|
||
|
|
"step": 245
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.246,
|
||
|
|
"grad_norm": 1.6328125,
|
||
|
|
"learning_rate": 0.00018746197071393958,
|
||
|
|
"loss": 2.9298,
|
||
|
|
"step": 246
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.247,
|
||
|
|
"grad_norm": 1.171875,
|
||
|
|
"learning_rate": 0.00018729220772698097,
|
||
|
|
"loss": 3.0326,
|
||
|
|
"step": 247
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.248,
|
||
|
|
"grad_norm": 1.0546875,
|
||
|
|
"learning_rate": 0.00018712138111201895,
|
||
|
|
"loss": 3.0219,
|
||
|
|
"step": 248
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.249,
|
||
|
|
"grad_norm": 1.078125,
|
||
|
|
"learning_rate": 0.0001869494929505219,
|
||
|
|
"loss": 3.0315,
|
||
|
|
"step": 249
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.25,
|
||
|
|
"grad_norm": 11.5,
|
||
|
|
"learning_rate": 0.00018677654533689287,
|
||
|
|
"loss": 2.9556,
|
||
|
|
"step": 250
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.25,
|
||
|
|
"eval_loss": 3.0629920959472656,
|
||
|
|
"eval_runtime": 31.7573,
|
||
|
|
"eval_samples_per_second": 15.744,
|
||
|
|
"eval_steps_per_second": 2.645,
|
||
|
|
"step": 250
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.251,
|
||
|
|
"grad_norm": 1.953125,
|
||
|
|
"learning_rate": 0.00018660254037844388,
|
||
|
|
"loss": 3.0847,
|
||
|
|
"step": 251
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.252,
|
||
|
|
"grad_norm": 1.328125,
|
||
|
|
"learning_rate": 0.0001864274801953705,
|
||
|
|
"loss": 3.0586,
|
||
|
|
"step": 252
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.253,
|
||
|
|
"grad_norm": 1.5390625,
|
||
|
|
"learning_rate": 0.00018625136692072575,
|
||
|
|
"loss": 3.1137,
|
||
|
|
"step": 253
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.254,
|
||
|
|
"grad_norm": 2.4375,
|
||
|
|
"learning_rate": 0.0001860742027003944,
|
||
|
|
"loss": 3.021,
|
||
|
|
"step": 254
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.255,
|
||
|
|
"grad_norm": 1.4375,
|
||
|
|
"learning_rate": 0.00018589598969306645,
|
||
|
|
"loss": 3.0077,
|
||
|
|
"step": 255
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.256,
|
||
|
|
"grad_norm": 1.4453125,
|
||
|
|
"learning_rate": 0.00018571673007021123,
|
||
|
|
"loss": 3.0533,
|
||
|
|
"step": 256
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.257,
|
||
|
|
"grad_norm": 1.453125,
|
||
|
|
"learning_rate": 0.00018553642601605068,
|
||
|
|
"loss": 3.038,
|
||
|
|
"step": 257
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.258,
|
||
|
|
"grad_norm": 2.125,
|
||
|
|
"learning_rate": 0.00018535507972753274,
|
||
|
|
"loss": 2.9965,
|
||
|
|
"step": 258
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.259,
|
||
|
|
"grad_norm": 1.34375,
|
||
|
|
"learning_rate": 0.00018517269341430476,
|
||
|
|
"loss": 3.0632,
|
||
|
|
"step": 259
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.26,
|
||
|
|
"grad_norm": 1.1640625,
|
||
|
|
"learning_rate": 0.00018498926929868642,
|
||
|
|
"loss": 3.1137,
|
||
|
|
"step": 260
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.261,
|
||
|
|
"grad_norm": 23.75,
|
||
|
|
"learning_rate": 0.0001848048096156426,
|
||
|
|
"loss": 2.9885,
|
||
|
|
"step": 261
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.262,
|
||
|
|
"grad_norm": 3.75,
|
||
|
|
"learning_rate": 0.00018461931661275643,
|
||
|
|
"loss": 3.12,
|
||
|
|
"step": 262
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.263,
|
||
|
|
"grad_norm": 2.0625,
|
||
|
|
"learning_rate": 0.00018443279255020152,
|
||
|
|
"loss": 3.1726,
|
||
|
|
"step": 263
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.264,
|
||
|
|
"grad_norm": 1.921875,
|
||
|
|
"learning_rate": 0.00018424523970071477,
|
||
|
|
"loss": 3.2129,
|
||
|
|
"step": 264
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.265,
|
||
|
|
"grad_norm": 1.71875,
|
||
|
|
"learning_rate": 0.00018405666034956844,
|
||
|
|
"loss": 2.9784,
|
||
|
|
"step": 265
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.266,
|
||
|
|
"grad_norm": 71.0,
|
||
|
|
"learning_rate": 0.00018386705679454242,
|
||
|
|
"loss": 3.0557,
|
||
|
|
"step": 266
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.267,
|
||
|
|
"grad_norm": 2.65625,
|
||
|
|
"learning_rate": 0.00018367643134589617,
|
||
|
|
"loss": 3.1084,
|
||
|
|
"step": 267
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.268,
|
||
|
|
"grad_norm": 1.8984375,
|
||
|
|
"learning_rate": 0.00018348478632634066,
|
||
|
|
"loss": 3.0229,
|
||
|
|
"step": 268
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.269,
|
||
|
|
"grad_norm": 3.71875,
|
||
|
|
"learning_rate": 0.00018329212407100994,
|
||
|
|
"loss": 2.9938,
|
||
|
|
"step": 269
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.27,
|
||
|
|
"grad_norm": 2.84375,
|
||
|
|
"learning_rate": 0.00018309844692743283,
|
||
|
|
"loss": 3.029,
|
||
|
|
"step": 270
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.271,
|
||
|
|
"grad_norm": 2.015625,
|
||
|
|
"learning_rate": 0.00018290375725550417,
|
||
|
|
"loss": 2.9886,
|
||
|
|
"step": 271
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.272,
|
||
|
|
"grad_norm": 2.015625,
|
||
|
|
"learning_rate": 0.00018270805742745617,
|
||
|
|
"loss": 3.0823,
|
||
|
|
"step": 272
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.273,
|
||
|
|
"grad_norm": 3.28125,
|
||
|
|
"learning_rate": 0.00018251134982782952,
|
||
|
|
"loss": 3.0021,
|
||
|
|
"step": 273
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.274,
|
||
|
|
"grad_norm": 2.203125,
|
||
|
|
"learning_rate": 0.0001823136368534442,
|
||
|
|
"loss": 2.9724,
|
||
|
|
"step": 274
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.275,
|
||
|
|
"grad_norm": 1.390625,
|
||
|
|
"learning_rate": 0.00018211492091337042,
|
||
|
|
"loss": 2.8975,
|
||
|
|
"step": 275
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.276,
|
||
|
|
"grad_norm": 1.8125,
|
||
|
|
"learning_rate": 0.0001819152044288992,
|
||
|
|
"loss": 2.9522,
|
||
|
|
"step": 276
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.277,
|
||
|
|
"grad_norm": 1.4921875,
|
||
|
|
"learning_rate": 0.00018171448983351284,
|
||
|
|
"loss": 2.9637,
|
||
|
|
"step": 277
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.278,
|
||
|
|
"grad_norm": 1.984375,
|
||
|
|
"learning_rate": 0.00018151277957285543,
|
||
|
|
"loss": 3.0239,
|
||
|
|
"step": 278
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.279,
|
||
|
|
"grad_norm": 1.59375,
|
||
|
|
"learning_rate": 0.00018131007610470276,
|
||
|
|
"loss": 2.9014,
|
||
|
|
"step": 279
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.28,
|
||
|
|
"grad_norm": 1.8828125,
|
||
|
|
"learning_rate": 0.00018110638189893267,
|
||
|
|
"loss": 3.0008,
|
||
|
|
"step": 280
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.281,
|
||
|
|
"grad_norm": 1.4765625,
|
||
|
|
"learning_rate": 0.00018090169943749476,
|
||
|
|
"loss": 3.1132,
|
||
|
|
"step": 281
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.282,
|
||
|
|
"grad_norm": 5.09375,
|
||
|
|
"learning_rate": 0.00018069603121438022,
|
||
|
|
"loss": 3.0337,
|
||
|
|
"step": 282
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.283,
|
||
|
|
"grad_norm": 2.765625,
|
||
|
|
"learning_rate": 0.0001804893797355914,
|
||
|
|
"loss": 3.1402,
|
||
|
|
"step": 283
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.284,
|
||
|
|
"grad_norm": 1.3984375,
|
||
|
|
"learning_rate": 0.00018028174751911146,
|
||
|
|
"loss": 2.9044,
|
||
|
|
"step": 284
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.285,
|
||
|
|
"grad_norm": 6.46875,
|
||
|
|
"learning_rate": 0.00018007313709487334,
|
||
|
|
"loss": 3.0565,
|
||
|
|
"step": 285
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.286,
|
||
|
|
"grad_norm": 2.65625,
|
||
|
|
"learning_rate": 0.00017986355100472928,
|
||
|
|
"loss": 3.1032,
|
||
|
|
"step": 286
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.287,
|
||
|
|
"grad_norm": 3.296875,
|
||
|
|
"learning_rate": 0.00017965299180241963,
|
||
|
|
"loss": 2.9643,
|
||
|
|
"step": 287
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.288,
|
||
|
|
"grad_norm": 2.078125,
|
||
|
|
"learning_rate": 0.00017944146205354182,
|
||
|
|
"loss": 3.0511,
|
||
|
|
"step": 288
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.289,
|
||
|
|
"grad_norm": 1.6328125,
|
||
|
|
"learning_rate": 0.00017922896433551907,
|
||
|
|
"loss": 2.8223,
|
||
|
|
"step": 289
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.29,
|
||
|
|
"grad_norm": 1.359375,
|
||
|
|
"learning_rate": 0.00017901550123756906,
|
||
|
|
"loss": 2.9322,
|
||
|
|
"step": 290
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.291,
|
||
|
|
"grad_norm": 13.0,
|
||
|
|
"learning_rate": 0.00017880107536067218,
|
||
|
|
"loss": 3.054,
|
||
|
|
"step": 291
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.292,
|
||
|
|
"grad_norm": 1.7265625,
|
||
|
|
"learning_rate": 0.0001785856893175402,
|
||
|
|
"loss": 3.0824,
|
||
|
|
"step": 292
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.293,
|
||
|
|
"grad_norm": 1.21875,
|
||
|
|
"learning_rate": 0.000178369345732584,
|
||
|
|
"loss": 3.0968,
|
||
|
|
"step": 293
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.294,
|
||
|
|
"grad_norm": 9.25,
|
||
|
|
"learning_rate": 0.00017815204724188187,
|
||
|
|
"loss": 3.138,
|
||
|
|
"step": 294
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.295,
|
||
|
|
"grad_norm": 1.703125,
|
||
|
|
"learning_rate": 0.00017793379649314744,
|
||
|
|
"loss": 3.0392,
|
||
|
|
"step": 295
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.296,
|
||
|
|
"grad_norm": 20.625,
|
||
|
|
"learning_rate": 0.0001777145961456971,
|
||
|
|
"loss": 2.9701,
|
||
|
|
"step": 296
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.297,
|
||
|
|
"grad_norm": 12.0,
|
||
|
|
"learning_rate": 0.00017749444887041799,
|
||
|
|
"loss": 3.0403,
|
||
|
|
"step": 297
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.298,
|
||
|
|
"grad_norm": 2.125,
|
||
|
|
"learning_rate": 0.00017727335734973512,
|
||
|
|
"loss": 2.8949,
|
||
|
|
"step": 298
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.299,
|
||
|
|
"grad_norm": 25.875,
|
||
|
|
"learning_rate": 0.00017705132427757895,
|
||
|
|
"loss": 3.0195,
|
||
|
|
"step": 299
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.3,
|
||
|
|
"grad_norm": 18.25,
|
||
|
|
"learning_rate": 0.00017682835235935236,
|
||
|
|
"loss": 2.9629,
|
||
|
|
"step": 300
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.301,
|
||
|
|
"grad_norm": 3.0,
|
||
|
|
"learning_rate": 0.0001766044443118978,
|
||
|
|
"loss": 3.0801,
|
||
|
|
"step": 301
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.302,
|
||
|
|
"grad_norm": 4.9375,
|
||
|
|
"learning_rate": 0.00017637960286346425,
|
||
|
|
"loss": 2.9368,
|
||
|
|
"step": 302
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.303,
|
||
|
|
"grad_norm": 2.203125,
|
||
|
|
"learning_rate": 0.0001761538307536737,
|
||
|
|
"loss": 2.9964,
|
||
|
|
"step": 303
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.304,
|
||
|
|
"grad_norm": 0.984375,
|
||
|
|
"learning_rate": 0.00017592713073348807,
|
||
|
|
"loss": 2.9558,
|
||
|
|
"step": 304
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.305,
|
||
|
|
"grad_norm": 1.1640625,
|
||
|
|
"learning_rate": 0.00017569950556517566,
|
||
|
|
"loss": 2.9006,
|
||
|
|
"step": 305
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.306,
|
||
|
|
"grad_norm": 1.4140625,
|
||
|
|
"learning_rate": 0.00017547095802227723,
|
||
|
|
"loss": 3.0037,
|
||
|
|
"step": 306
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.307,
|
||
|
|
"grad_norm": 1.0,
|
||
|
|
"learning_rate": 0.00017524149088957245,
|
||
|
|
"loss": 2.8517,
|
||
|
|
"step": 307
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.308,
|
||
|
|
"grad_norm": 0.92578125,
|
||
|
|
"learning_rate": 0.00017501110696304596,
|
||
|
|
"loss": 2.9163,
|
||
|
|
"step": 308
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.309,
|
||
|
|
"grad_norm": 0.98046875,
|
||
|
|
"learning_rate": 0.0001747798090498532,
|
||
|
|
"loss": 2.9563,
|
||
|
|
"step": 309
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.31,
|
||
|
|
"grad_norm": 0.8984375,
|
||
|
|
"learning_rate": 0.00017454759996828623,
|
||
|
|
"loss": 2.8861,
|
||
|
|
"step": 310
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.311,
|
||
|
|
"grad_norm": 0.94140625,
|
||
|
|
"learning_rate": 0.00017431448254773944,
|
||
|
|
"loss": 3.0871,
|
||
|
|
"step": 311
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.312,
|
||
|
|
"grad_norm": 0.98046875,
|
||
|
|
"learning_rate": 0.000174080459628675,
|
||
|
|
"loss": 2.8852,
|
||
|
|
"step": 312
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.313,
|
||
|
|
"grad_norm": 26.0,
|
||
|
|
"learning_rate": 0.00017384553406258842,
|
||
|
|
"loss": 3.0938,
|
||
|
|
"step": 313
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.314,
|
||
|
|
"grad_norm": 1.7265625,
|
||
|
|
"learning_rate": 0.00017360970871197346,
|
||
|
|
"loss": 3.0905,
|
||
|
|
"step": 314
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.315,
|
||
|
|
"grad_norm": 1.140625,
|
||
|
|
"learning_rate": 0.00017337298645028764,
|
||
|
|
"loss": 3.0793,
|
||
|
|
"step": 315
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.316,
|
||
|
|
"grad_norm": 1.25,
|
||
|
|
"learning_rate": 0.00017313537016191706,
|
||
|
|
"loss": 3.0879,
|
||
|
|
"step": 316
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.317,
|
||
|
|
"grad_norm": 1.1875,
|
||
|
|
"learning_rate": 0.00017289686274214118,
|
||
|
|
"loss": 3.1008,
|
||
|
|
"step": 317
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.318,
|
||
|
|
"grad_norm": 1.1015625,
|
||
|
|
"learning_rate": 0.0001726574670970976,
|
||
|
|
"loss": 3.0737,
|
||
|
|
"step": 318
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.319,
|
||
|
|
"grad_norm": 1.0859375,
|
||
|
|
"learning_rate": 0.00017241718614374678,
|
||
|
|
"loss": 2.9513,
|
||
|
|
"step": 319
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.32,
|
||
|
|
"grad_norm": 1.046875,
|
||
|
|
"learning_rate": 0.00017217602280983623,
|
||
|
|
"loss": 3.1167,
|
||
|
|
"step": 320
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.321,
|
||
|
|
"grad_norm": 1.0,
|
||
|
|
"learning_rate": 0.0001719339800338651,
|
||
|
|
"loss": 3.0823,
|
||
|
|
"step": 321
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.322,
|
||
|
|
"grad_norm": 1.03125,
|
||
|
|
"learning_rate": 0.0001716910607650483,
|
||
|
|
"loss": 2.9707,
|
||
|
|
"step": 322
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.323,
|
||
|
|
"grad_norm": 0.9921875,
|
||
|
|
"learning_rate": 0.00017144726796328034,
|
||
|
|
"loss": 3.0313,
|
||
|
|
"step": 323
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.324,
|
||
|
|
"grad_norm": 0.93359375,
|
||
|
|
"learning_rate": 0.00017120260459909967,
|
||
|
|
"loss": 2.9786,
|
||
|
|
"step": 324
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.325,
|
||
|
|
"grad_norm": 1.0703125,
|
||
|
|
"learning_rate": 0.0001709570736536521,
|
||
|
|
"loss": 2.9752,
|
||
|
|
"step": 325
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.326,
|
||
|
|
"grad_norm": 0.91796875,
|
||
|
|
"learning_rate": 0.00017071067811865476,
|
||
|
|
"loss": 3.0147,
|
||
|
|
"step": 326
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.327,
|
||
|
|
"grad_norm": 1.1328125,
|
||
|
|
"learning_rate": 0.00017046342099635948,
|
||
|
|
"loss": 3.0414,
|
||
|
|
"step": 327
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.328,
|
||
|
|
"grad_norm": 0.94921875,
|
||
|
|
"learning_rate": 0.00017021530529951625,
|
||
|
|
"loss": 3.0018,
|
||
|
|
"step": 328
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.329,
|
||
|
|
"grad_norm": 1.265625,
|
||
|
|
"learning_rate": 0.00016996633405133655,
|
||
|
|
"loss": 3.0249,
|
||
|
|
"step": 329
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.33,
|
||
|
|
"grad_norm": 1.2109375,
|
||
|
|
"learning_rate": 0.00016971651028545648,
|
||
|
|
"loss": 2.9958,
|
||
|
|
"step": 330
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.331,
|
||
|
|
"grad_norm": 0.8984375,
|
||
|
|
"learning_rate": 0.00016946583704589973,
|
||
|
|
"loss": 2.9505,
|
||
|
|
"step": 331
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.332,
|
||
|
|
"grad_norm": 1.0703125,
|
||
|
|
"learning_rate": 0.0001692143173870407,
|
||
|
|
"loss": 2.9806,
|
||
|
|
"step": 332
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.333,
|
||
|
|
"grad_norm": 248.0,
|
||
|
|
"learning_rate": 0.000168961954373567,
|
||
|
|
"loss": 2.9853,
|
||
|
|
"step": 333
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.334,
|
||
|
|
"grad_norm": 1.75,
|
||
|
|
"learning_rate": 0.0001687087510804423,
|
||
|
|
"loss": 2.9122,
|
||
|
|
"step": 334
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.335,
|
||
|
|
"grad_norm": 1.0625,
|
||
|
|
"learning_rate": 0.00016845471059286887,
|
||
|
|
"loss": 2.9686,
|
||
|
|
"step": 335
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.336,
|
||
|
|
"grad_norm": 1.4375,
|
||
|
|
"learning_rate": 0.00016819983600624986,
|
||
|
|
"loss": 3.0128,
|
||
|
|
"step": 336
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.337,
|
||
|
|
"grad_norm": 1.4609375,
|
||
|
|
"learning_rate": 0.00016794413042615168,
|
||
|
|
"loss": 2.9692,
|
||
|
|
"step": 337
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.338,
|
||
|
|
"grad_norm": 1.4140625,
|
||
|
|
"learning_rate": 0.00016768759696826608,
|
||
|
|
"loss": 2.9639,
|
||
|
|
"step": 338
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.339,
|
||
|
|
"grad_norm": 1.140625,
|
||
|
|
"learning_rate": 0.00016743023875837233,
|
||
|
|
"loss": 2.8997,
|
||
|
|
"step": 339
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.34,
|
||
|
|
"grad_norm": 1.2421875,
|
||
|
|
"learning_rate": 0.00016717205893229903,
|
||
|
|
"loss": 2.9825,
|
||
|
|
"step": 340
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.341,
|
||
|
|
"grad_norm": 1.1171875,
|
||
|
|
"learning_rate": 0.00016691306063588583,
|
||
|
|
"loss": 2.9442,
|
||
|
|
"step": 341
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.342,
|
||
|
|
"grad_norm": 1.1171875,
|
||
|
|
"learning_rate": 0.00016665324702494524,
|
||
|
|
"loss": 2.9611,
|
||
|
|
"step": 342
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.343,
|
||
|
|
"grad_norm": 1.0,
|
||
|
|
"learning_rate": 0.00016639262126522418,
|
||
|
|
"loss": 3.0123,
|
||
|
|
"step": 343
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.344,
|
||
|
|
"grad_norm": 1.6953125,
|
||
|
|
"learning_rate": 0.00016613118653236518,
|
||
|
|
"loss": 2.9498,
|
||
|
|
"step": 344
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.345,
|
||
|
|
"grad_norm": 1.15625,
|
||
|
|
"learning_rate": 0.00016586894601186805,
|
||
|
|
"loss": 2.8588,
|
||
|
|
"step": 345
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.346,
|
||
|
|
"grad_norm": 0.9609375,
|
||
|
|
"learning_rate": 0.00016560590289905073,
|
||
|
|
"loss": 3.0459,
|
||
|
|
"step": 346
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.347,
|
||
|
|
"grad_norm": 1.0703125,
|
||
|
|
"learning_rate": 0.00016534206039901057,
|
||
|
|
"loss": 2.8838,
|
||
|
|
"step": 347
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.348,
|
||
|
|
"grad_norm": 0.95703125,
|
||
|
|
"learning_rate": 0.0001650774217265851,
|
||
|
|
"loss": 2.9694,
|
||
|
|
"step": 348
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.349,
|
||
|
|
"grad_norm": 0.92578125,
|
||
|
|
"learning_rate": 0.0001648119901063131,
|
||
|
|
"loss": 2.9102,
|
||
|
|
"step": 349
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.35,
|
||
|
|
"grad_norm": 0.921875,
|
||
|
|
"learning_rate": 0.00016454576877239507,
|
||
|
|
"loss": 2.8904,
|
||
|
|
"step": 350
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.351,
|
||
|
|
"grad_norm": 0.85546875,
|
||
|
|
"learning_rate": 0.00016427876096865394,
|
||
|
|
"loss": 2.8798,
|
||
|
|
"step": 351
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.352,
|
||
|
|
"grad_norm": 0.93359375,
|
||
|
|
"learning_rate": 0.00016401096994849557,
|
||
|
|
"loss": 2.9685,
|
||
|
|
"step": 352
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.353,
|
||
|
|
"grad_norm": 0.890625,
|
||
|
|
"learning_rate": 0.000163742398974869,
|
||
|
|
"loss": 2.9498,
|
||
|
|
"step": 353
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.354,
|
||
|
|
"grad_norm": 0.93359375,
|
||
|
|
"learning_rate": 0.00016347305132022677,
|
||
|
|
"loss": 2.9719,
|
||
|
|
"step": 354
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.355,
|
||
|
|
"grad_norm": 0.87890625,
|
||
|
|
"learning_rate": 0.0001632029302664851,
|
||
|
|
"loss": 2.9964,
|
||
|
|
"step": 355
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.356,
|
||
|
|
"grad_norm": 0.921875,
|
||
|
|
"learning_rate": 0.00016293203910498376,
|
||
|
|
"loss": 2.9672,
|
||
|
|
"step": 356
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.357,
|
||
|
|
"grad_norm": 0.87109375,
|
||
|
|
"learning_rate": 0.00016266038113644607,
|
||
|
|
"loss": 2.944,
|
||
|
|
"step": 357
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.358,
|
||
|
|
"grad_norm": 0.87890625,
|
||
|
|
"learning_rate": 0.00016238795967093864,
|
||
|
|
"loss": 2.928,
|
||
|
|
"step": 358
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.359,
|
||
|
|
"grad_norm": 0.91015625,
|
||
|
|
"learning_rate": 0.00016211477802783103,
|
||
|
|
"loss": 2.8984,
|
||
|
|
"step": 359
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.36,
|
||
|
|
"grad_norm": 1.6640625,
|
||
|
|
"learning_rate": 0.0001618408395357554,
|
||
|
|
"loss": 2.807,
|
||
|
|
"step": 360
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.361,
|
||
|
|
"grad_norm": 0.98828125,
|
||
|
|
"learning_rate": 0.0001615661475325658,
|
||
|
|
"loss": 2.9199,
|
||
|
|
"step": 361
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.362,
|
||
|
|
"grad_norm": 1.1171875,
|
||
|
|
"learning_rate": 0.00016129070536529766,
|
||
|
|
"loss": 2.9442,
|
||
|
|
"step": 362
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.363,
|
||
|
|
"grad_norm": 1.1484375,
|
||
|
|
"learning_rate": 0.0001610145163901268,
|
||
|
|
"loss": 2.8548,
|
||
|
|
"step": 363
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.364,
|
||
|
|
"grad_norm": 0.8671875,
|
||
|
|
"learning_rate": 0.00016073758397232868,
|
||
|
|
"loss": 2.9272,
|
||
|
|
"step": 364
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.365,
|
||
|
|
"grad_norm": 0.890625,
|
||
|
|
"learning_rate": 0.0001604599114862375,
|
||
|
|
"loss": 2.9774,
|
||
|
|
"step": 365
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.366,
|
||
|
|
"grad_norm": 0.9609375,
|
||
|
|
"learning_rate": 0.00016018150231520486,
|
||
|
|
"loss": 2.8988,
|
||
|
|
"step": 366
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.367,
|
||
|
|
"grad_norm": 0.8984375,
|
||
|
|
"learning_rate": 0.0001599023598515586,
|
||
|
|
"loss": 2.8573,
|
||
|
|
"step": 367
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.368,
|
||
|
|
"grad_norm": 1.4609375,
|
||
|
|
"learning_rate": 0.0001596224874965616,
|
||
|
|
"loss": 2.86,
|
||
|
|
"step": 368
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.369,
|
||
|
|
"grad_norm": 0.890625,
|
||
|
|
"learning_rate": 0.00015934188866037016,
|
||
|
|
"loss": 2.9432,
|
||
|
|
"step": 369
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.37,
|
||
|
|
"grad_norm": 0.875,
|
||
|
|
"learning_rate": 0.00015906056676199255,
|
||
|
|
"loss": 2.8969,
|
||
|
|
"step": 370
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.371,
|
||
|
|
"grad_norm": 0.91015625,
|
||
|
|
"learning_rate": 0.00015877852522924732,
|
||
|
|
"loss": 2.968,
|
||
|
|
"step": 371
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.372,
|
||
|
|
"grad_norm": 0.86328125,
|
||
|
|
"learning_rate": 0.00015849576749872157,
|
||
|
|
"loss": 2.885,
|
||
|
|
"step": 372
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.373,
|
||
|
|
"grad_norm": 0.82421875,
|
||
|
|
"learning_rate": 0.00015821229701572896,
|
||
|
|
"loss": 2.9352,
|
||
|
|
"step": 373
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.374,
|
||
|
|
"grad_norm": 0.87109375,
|
||
|
|
"learning_rate": 0.0001579281172342679,
|
||
|
|
"loss": 2.7847,
|
||
|
|
"step": 374
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.375,
|
||
|
|
"grad_norm": 1.2109375,
|
||
|
|
"learning_rate": 0.00015764323161697935,
|
||
|
|
"loss": 2.9347,
|
||
|
|
"step": 375
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.376,
|
||
|
|
"grad_norm": 0.98046875,
|
||
|
|
"learning_rate": 0.0001573576436351046,
|
||
|
|
"loss": 2.8265,
|
||
|
|
"step": 376
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.377,
|
||
|
|
"grad_norm": 0.921875,
|
||
|
|
"learning_rate": 0.0001570713567684432,
|
||
|
|
"loss": 2.8173,
|
||
|
|
"step": 377
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.378,
|
||
|
|
"grad_norm": 0.9921875,
|
||
|
|
"learning_rate": 0.00015678437450531013,
|
||
|
|
"loss": 2.8425,
|
||
|
|
"step": 378
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.379,
|
||
|
|
"grad_norm": 0.91015625,
|
||
|
|
"learning_rate": 0.0001564967003424938,
|
||
|
|
"loss": 2.7723,
|
||
|
|
"step": 379
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.38,
|
||
|
|
"grad_norm": 0.859375,
|
||
|
|
"learning_rate": 0.00015620833778521307,
|
||
|
|
"loss": 2.9031,
|
||
|
|
"step": 380
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.381,
|
||
|
|
"grad_norm": 0.85546875,
|
||
|
|
"learning_rate": 0.0001559192903470747,
|
||
|
|
"loss": 2.9154,
|
||
|
|
"step": 381
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.382,
|
||
|
|
"grad_norm": 0.86328125,
|
||
|
|
"learning_rate": 0.0001556295615500305,
|
||
|
|
"loss": 2.8226,
|
||
|
|
"step": 382
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.383,
|
||
|
|
"grad_norm": 1.015625,
|
||
|
|
"learning_rate": 0.00015533915492433443,
|
||
|
|
"loss": 2.8911,
|
||
|
|
"step": 383
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.384,
|
||
|
|
"grad_norm": 0.8828125,
|
||
|
|
"learning_rate": 0.00015504807400849958,
|
||
|
|
"loss": 2.8114,
|
||
|
|
"step": 384
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.385,
|
||
|
|
"grad_norm": 0.890625,
|
||
|
|
"learning_rate": 0.00015475632234925504,
|
||
|
|
"loss": 2.8762,
|
||
|
|
"step": 385
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.386,
|
||
|
|
"grad_norm": 0.828125,
|
||
|
|
"learning_rate": 0.00015446390350150273,
|
||
|
|
"loss": 2.8918,
|
||
|
|
"step": 386
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.387,
|
||
|
|
"grad_norm": 0.9296875,
|
||
|
|
"learning_rate": 0.000154170821028274,
|
||
|
|
"loss": 2.9433,
|
||
|
|
"step": 387
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.388,
|
||
|
|
"grad_norm": 0.84375,
|
||
|
|
"learning_rate": 0.0001538770785006863,
|
||
|
|
"loss": 2.871,
|
||
|
|
"step": 388
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.389,
|
||
|
|
"grad_norm": 0.90625,
|
||
|
|
"learning_rate": 0.00015358267949789966,
|
||
|
|
"loss": 2.8932,
|
||
|
|
"step": 389
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.39,
|
||
|
|
"grad_norm": 0.86328125,
|
||
|
|
"learning_rate": 0.000153287627607073,
|
||
|
|
"loss": 2.8251,
|
||
|
|
"step": 390
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.391,
|
||
|
|
"grad_norm": 1.21875,
|
||
|
|
"learning_rate": 0.0001529919264233205,
|
||
|
|
"loss": 2.9062,
|
||
|
|
"step": 391
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.392,
|
||
|
|
"grad_norm": 1.0625,
|
||
|
|
"learning_rate": 0.00015269557954966778,
|
||
|
|
"loss": 2.8995,
|
||
|
|
"step": 392
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.393,
|
||
|
|
"grad_norm": 0.796875,
|
||
|
|
"learning_rate": 0.00015239859059700794,
|
||
|
|
"loss": 2.7726,
|
||
|
|
"step": 393
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.394,
|
||
|
|
"grad_norm": 1.0625,
|
||
|
|
"learning_rate": 0.00015210096318405767,
|
||
|
|
"loss": 2.7649,
|
||
|
|
"step": 394
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.395,
|
||
|
|
"grad_norm": 0.92578125,
|
||
|
|
"learning_rate": 0.00015180270093731303,
|
||
|
|
"loss": 2.778,
|
||
|
|
"step": 395
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.396,
|
||
|
|
"grad_norm": 0.94921875,
|
||
|
|
"learning_rate": 0.00015150380749100545,
|
||
|
|
"loss": 2.8322,
|
||
|
|
"step": 396
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.397,
|
||
|
|
"grad_norm": 0.91796875,
|
||
|
|
"learning_rate": 0.00015120428648705717,
|
||
|
|
"loss": 2.8831,
|
||
|
|
"step": 397
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.398,
|
||
|
|
"grad_norm": 0.8671875,
|
||
|
|
"learning_rate": 0.00015090414157503714,
|
||
|
|
"loss": 2.8862,
|
||
|
|
"step": 398
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.399,
|
||
|
|
"grad_norm": 0.84765625,
|
||
|
|
"learning_rate": 0.00015060337641211637,
|
||
|
|
"loss": 2.8219,
|
||
|
|
"step": 399
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.4,
|
||
|
|
"grad_norm": 0.7890625,
|
||
|
|
"learning_rate": 0.00015030199466302353,
|
||
|
|
"loss": 2.8894,
|
||
|
|
"step": 400
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.401,
|
||
|
|
"grad_norm": 0.83984375,
|
||
|
|
"learning_rate": 0.00015000000000000001,
|
||
|
|
"loss": 2.8594,
|
||
|
|
"step": 401
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.402,
|
||
|
|
"grad_norm": 0.875,
|
||
|
|
"learning_rate": 0.00014969739610275556,
|
||
|
|
"loss": 2.7943,
|
||
|
|
"step": 402
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.403,
|
||
|
|
"grad_norm": 0.890625,
|
||
|
|
"learning_rate": 0.0001493941866584231,
|
||
|
|
"loss": 2.9105,
|
||
|
|
"step": 403
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.404,
|
||
|
|
"grad_norm": 1.3046875,
|
||
|
|
"learning_rate": 0.00014909037536151409,
|
||
|
|
"loss": 2.9341,
|
||
|
|
"step": 404
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.405,
|
||
|
|
"grad_norm": 0.984375,
|
||
|
|
"learning_rate": 0.0001487859659138733,
|
||
|
|
"loss": 2.8726,
|
||
|
|
"step": 405
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.406,
|
||
|
|
"grad_norm": 0.796875,
|
||
|
|
"learning_rate": 0.00014848096202463372,
|
||
|
|
"loss": 2.9266,
|
||
|
|
"step": 406
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.407,
|
||
|
|
"grad_norm": 12.5,
|
||
|
|
"learning_rate": 0.00014817536741017152,
|
||
|
|
"loss": 2.8098,
|
||
|
|
"step": 407
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.408,
|
||
|
|
"grad_norm": 1.21875,
|
||
|
|
"learning_rate": 0.0001478691857940607,
|
||
|
|
"loss": 2.8836,
|
||
|
|
"step": 408
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.409,
|
||
|
|
"grad_norm": 1.1328125,
|
||
|
|
"learning_rate": 0.00014756242090702756,
|
||
|
|
"loss": 2.8176,
|
||
|
|
"step": 409
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.41,
|
||
|
|
"grad_norm": 1.265625,
|
||
|
|
"learning_rate": 0.00014725507648690543,
|
||
|
|
"loss": 2.8858,
|
||
|
|
"step": 410
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.411,
|
||
|
|
"grad_norm": 1.1328125,
|
||
|
|
"learning_rate": 0.00014694715627858908,
|
||
|
|
"loss": 2.8898,
|
||
|
|
"step": 411
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.412,
|
||
|
|
"grad_norm": 1.078125,
|
||
|
|
"learning_rate": 0.00014663866403398913,
|
||
|
|
"loss": 2.8529,
|
||
|
|
"step": 412
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.413,
|
||
|
|
"grad_norm": 1.0625,
|
||
|
|
"learning_rate": 0.00014632960351198618,
|
||
|
|
"loss": 2.8151,
|
||
|
|
"step": 413
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.414,
|
||
|
|
"grad_norm": 1.0078125,
|
||
|
|
"learning_rate": 0.00014601997847838518,
|
||
|
|
"loss": 2.9201,
|
||
|
|
"step": 414
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.415,
|
||
|
|
"grad_norm": 1.0078125,
|
||
|
|
"learning_rate": 0.00014570979270586945,
|
||
|
|
"loss": 2.7901,
|
||
|
|
"step": 415
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.416,
|
||
|
|
"grad_norm": 1.1015625,
|
||
|
|
"learning_rate": 0.00014539904997395468,
|
||
|
|
"loss": 2.8675,
|
||
|
|
"step": 416
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.417,
|
||
|
|
"grad_norm": 0.9453125,
|
||
|
|
"learning_rate": 0.00014508775406894307,
|
||
|
|
"loss": 2.7743,
|
||
|
|
"step": 417
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.418,
|
||
|
|
"grad_norm": 0.98046875,
|
||
|
|
"learning_rate": 0.00014477590878387696,
|
||
|
|
"loss": 2.8796,
|
||
|
|
"step": 418
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.419,
|
||
|
|
"grad_norm": 0.94921875,
|
||
|
|
"learning_rate": 0.00014446351791849276,
|
||
|
|
"loss": 2.9855,
|
||
|
|
"step": 419
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.42,
|
||
|
|
"grad_norm": 0.95703125,
|
||
|
|
"learning_rate": 0.00014415058527917452,
|
||
|
|
"loss": 2.8645,
|
||
|
|
"step": 420
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.421,
|
||
|
|
"grad_norm": 0.91015625,
|
||
|
|
"learning_rate": 0.00014383711467890774,
|
||
|
|
"loss": 2.7821,
|
||
|
|
"step": 421
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.422,
|
||
|
|
"grad_norm": 5.53125,
|
||
|
|
"learning_rate": 0.00014352310993723277,
|
||
|
|
"loss": 2.9062,
|
||
|
|
"step": 422
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.423,
|
||
|
|
"grad_norm": 0.96484375,
|
||
|
|
"learning_rate": 0.00014320857488019824,
|
||
|
|
"loss": 2.8564,
|
||
|
|
"step": 423
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.424,
|
||
|
|
"grad_norm": 0.84765625,
|
||
|
|
"learning_rate": 0.0001428935133403146,
|
||
|
|
"loss": 2.855,
|
||
|
|
"step": 424
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.425,
|
||
|
|
"grad_norm": 0.91796875,
|
||
|
|
"learning_rate": 0.00014257792915650728,
|
||
|
|
"loss": 2.907,
|
||
|
|
"step": 425
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.426,
|
||
|
|
"grad_norm": 1.0546875,
|
||
|
|
"learning_rate": 0.00014226182617406996,
|
||
|
|
"loss": 2.8238,
|
||
|
|
"step": 426
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.427,
|
||
|
|
"grad_norm": 1.09375,
|
||
|
|
"learning_rate": 0.00014194520824461771,
|
||
|
|
"loss": 2.8523,
|
||
|
|
"step": 427
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.428,
|
||
|
|
"grad_norm": 1.1015625,
|
||
|
|
"learning_rate": 0.00014162807922604012,
|
||
|
|
"loss": 2.91,
|
||
|
|
"step": 428
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.429,
|
||
|
|
"grad_norm": 1.2109375,
|
||
|
|
"learning_rate": 0.0001413104429824542,
|
||
|
|
"loss": 2.9161,
|
||
|
|
"step": 429
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.43,
|
||
|
|
"grad_norm": 0.8515625,
|
||
|
|
"learning_rate": 0.00014099230338415728,
|
||
|
|
"loss": 2.9304,
|
||
|
|
"step": 430
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.431,
|
||
|
|
"grad_norm": 0.75,
|
||
|
|
"learning_rate": 0.00014067366430758004,
|
||
|
|
"loss": 2.8266,
|
||
|
|
"step": 431
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.432,
|
||
|
|
"grad_norm": 0.8203125,
|
||
|
|
"learning_rate": 0.00014035452963523902,
|
||
|
|
"loss": 2.8303,
|
||
|
|
"step": 432
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.433,
|
||
|
|
"grad_norm": 0.7890625,
|
||
|
|
"learning_rate": 0.00014003490325568954,
|
||
|
|
"loss": 2.8311,
|
||
|
|
"step": 433
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.434,
|
||
|
|
"grad_norm": 0.7265625,
|
||
|
|
"learning_rate": 0.00013971478906347806,
|
||
|
|
"loss": 2.826,
|
||
|
|
"step": 434
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.435,
|
||
|
|
"grad_norm": 0.8828125,
|
||
|
|
"learning_rate": 0.00013939419095909512,
|
||
|
|
"loss": 2.6267,
|
||
|
|
"step": 435
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.436,
|
||
|
|
"grad_norm": 0.77734375,
|
||
|
|
"learning_rate": 0.00013907311284892736,
|
||
|
|
"loss": 2.7203,
|
||
|
|
"step": 436
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.437,
|
||
|
|
"grad_norm": 0.7421875,
|
||
|
|
"learning_rate": 0.0001387515586452103,
|
||
|
|
"loss": 2.8801,
|
||
|
|
"step": 437
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.438,
|
||
|
|
"grad_norm": 2.421875,
|
||
|
|
"learning_rate": 0.00013842953226598037,
|
||
|
|
"loss": 2.9568,
|
||
|
|
"step": 438
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.439,
|
||
|
|
"grad_norm": 0.9375,
|
||
|
|
"learning_rate": 0.00013810703763502744,
|
||
|
|
"loss": 2.758,
|
||
|
|
"step": 439
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.44,
|
||
|
|
"grad_norm": 1.3046875,
|
||
|
|
"learning_rate": 0.00013778407868184672,
|
||
|
|
"loss": 2.8374,
|
||
|
|
"step": 440
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.441,
|
||
|
|
"grad_norm": 0.91796875,
|
||
|
|
"learning_rate": 0.00013746065934159123,
|
||
|
|
"loss": 2.7894,
|
||
|
|
"step": 441
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.442,
|
||
|
|
"grad_norm": 0.87109375,
|
||
|
|
"learning_rate": 0.00013713678355502351,
|
||
|
|
"loss": 2.855,
|
||
|
|
"step": 442
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.443,
|
||
|
|
"grad_norm": 0.734375,
|
||
|
|
"learning_rate": 0.00013681245526846783,
|
||
|
|
"loss": 2.777,
|
||
|
|
"step": 443
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.444,
|
||
|
|
"grad_norm": 0.875,
|
||
|
|
"learning_rate": 0.00013648767843376196,
|
||
|
|
"loss": 2.8408,
|
||
|
|
"step": 444
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.445,
|
||
|
|
"grad_norm": 0.80078125,
|
||
|
|
"learning_rate": 0.00013616245700820922,
|
||
|
|
"loss": 2.8602,
|
||
|
|
"step": 445
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.446,
|
||
|
|
"grad_norm": 0.8359375,
|
||
|
|
"learning_rate": 0.00013583679495453,
|
||
|
|
"loss": 2.9452,
|
||
|
|
"step": 446
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.447,
|
||
|
|
"grad_norm": 0.83203125,
|
||
|
|
"learning_rate": 0.0001355106962408137,
|
||
|
|
"loss": 2.799,
|
||
|
|
"step": 447
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.448,
|
||
|
|
"grad_norm": 0.78125,
|
||
|
|
"learning_rate": 0.00013518416484047018,
|
||
|
|
"loss": 2.8164,
|
||
|
|
"step": 448
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.449,
|
||
|
|
"grad_norm": 0.78515625,
|
||
|
|
"learning_rate": 0.00013485720473218154,
|
||
|
|
"loss": 2.9159,
|
||
|
|
"step": 449
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.45,
|
||
|
|
"grad_norm": 0.77734375,
|
||
|
|
"learning_rate": 0.00013452981989985348,
|
||
|
|
"loss": 2.8874,
|
||
|
|
"step": 450
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.451,
|
||
|
|
"grad_norm": 0.984375,
|
||
|
|
"learning_rate": 0.00013420201433256689,
|
||
|
|
"loss": 2.8846,
|
||
|
|
"step": 451
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.452,
|
||
|
|
"grad_norm": 0.78515625,
|
||
|
|
"learning_rate": 0.00013387379202452917,
|
||
|
|
"loss": 2.8474,
|
||
|
|
"step": 452
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.453,
|
||
|
|
"grad_norm": 0.86328125,
|
||
|
|
"learning_rate": 0.00013354515697502553,
|
||
|
|
"loss": 2.8697,
|
||
|
|
"step": 453
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.454,
|
||
|
|
"grad_norm": 2.875,
|
||
|
|
"learning_rate": 0.00013321611318837032,
|
||
|
|
"loss": 2.8103,
|
||
|
|
"step": 454
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.455,
|
||
|
|
"grad_norm": 0.88671875,
|
||
|
|
"learning_rate": 0.00013288666467385833,
|
||
|
|
"loss": 2.7893,
|
||
|
|
"step": 455
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.456,
|
||
|
|
"grad_norm": 0.734375,
|
||
|
|
"learning_rate": 0.00013255681544571568,
|
||
|
|
"loss": 2.8302,
|
||
|
|
"step": 456
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.457,
|
||
|
|
"grad_norm": 0.76953125,
|
||
|
|
"learning_rate": 0.00013222656952305113,
|
||
|
|
"loss": 2.8056,
|
||
|
|
"step": 457
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.458,
|
||
|
|
"grad_norm": 0.7734375,
|
||
|
|
"learning_rate": 0.00013189593092980702,
|
||
|
|
"loss": 2.8463,
|
||
|
|
"step": 458
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.459,
|
||
|
|
"grad_norm": 0.75,
|
||
|
|
"learning_rate": 0.00013156490369471027,
|
||
|
|
"loss": 2.7264,
|
||
|
|
"step": 459
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.46,
|
||
|
|
"grad_norm": 0.75390625,
|
||
|
|
"learning_rate": 0.00013123349185122327,
|
||
|
|
"loss": 2.7938,
|
||
|
|
"step": 460
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.461,
|
||
|
|
"grad_norm": 0.76953125,
|
||
|
|
"learning_rate": 0.00013090169943749476,
|
||
|
|
"loss": 2.7834,
|
||
|
|
"step": 461
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.462,
|
||
|
|
"grad_norm": 0.7265625,
|
||
|
|
"learning_rate": 0.00013056953049631057,
|
||
|
|
"loss": 2.8079,
|
||
|
|
"step": 462
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.463,
|
||
|
|
"grad_norm": 0.73046875,
|
||
|
|
"learning_rate": 0.00013023698907504446,
|
||
|
|
"loss": 2.7598,
|
||
|
|
"step": 463
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.464,
|
||
|
|
"grad_norm": 0.71875,
|
||
|
|
"learning_rate": 0.00012990407922560868,
|
||
|
|
"loss": 2.8624,
|
||
|
|
"step": 464
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.465,
|
||
|
|
"grad_norm": 0.7421875,
|
||
|
|
"learning_rate": 0.00012957080500440468,
|
||
|
|
"loss": 2.8337,
|
||
|
|
"step": 465
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.466,
|
||
|
|
"grad_norm": 0.71484375,
|
||
|
|
"learning_rate": 0.00012923717047227368,
|
||
|
|
"loss": 2.8641,
|
||
|
|
"step": 466
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.467,
|
||
|
|
"grad_norm": 0.6953125,
|
||
|
|
"learning_rate": 0.00012890317969444716,
|
||
|
|
"loss": 2.832,
|
||
|
|
"step": 467
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.468,
|
||
|
|
"grad_norm": 0.765625,
|
||
|
|
"learning_rate": 0.00012856883674049736,
|
||
|
|
"loss": 2.9126,
|
||
|
|
"step": 468
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.469,
|
||
|
|
"grad_norm": 3.4375,
|
||
|
|
"learning_rate": 0.00012823414568428768,
|
||
|
|
"loss": 2.9386,
|
||
|
|
"step": 469
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.47,
|
||
|
|
"grad_norm": 0.81640625,
|
||
|
|
"learning_rate": 0.00012789911060392294,
|
||
|
|
"loss": 2.7886,
|
||
|
|
"step": 470
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.471,
|
||
|
|
"grad_norm": 0.7421875,
|
||
|
|
"learning_rate": 0.0001275637355816999,
|
||
|
|
"loss": 2.9174,
|
||
|
|
"step": 471
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.472,
|
||
|
|
"grad_norm": 0.7890625,
|
||
|
|
"learning_rate": 0.00012722802470405744,
|
||
|
|
"loss": 2.7964,
|
||
|
|
"step": 472
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.473,
|
||
|
|
"grad_norm": 0.83984375,
|
||
|
|
"learning_rate": 0.00012689198206152657,
|
||
|
|
"loss": 2.779,
|
||
|
|
"step": 473
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.474,
|
||
|
|
"grad_norm": 0.7421875,
|
||
|
|
"learning_rate": 0.00012655561174868088,
|
||
|
|
"loss": 2.8909,
|
||
|
|
"step": 474
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.475,
|
||
|
|
"grad_norm": 0.7890625,
|
||
|
|
"learning_rate": 0.00012621891786408648,
|
||
|
|
"loss": 2.8705,
|
||
|
|
"step": 475
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.476,
|
||
|
|
"grad_norm": 0.83203125,
|
||
|
|
"learning_rate": 0.00012588190451025207,
|
||
|
|
"loss": 2.9862,
|
||
|
|
"step": 476
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.477,
|
||
|
|
"grad_norm": 0.74609375,
|
||
|
|
"learning_rate": 0.00012554457579357905,
|
||
|
|
"loss": 2.8599,
|
||
|
|
"step": 477
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.478,
|
||
|
|
"grad_norm": 0.73828125,
|
||
|
|
"learning_rate": 0.0001252069358243114,
|
||
|
|
"loss": 2.9299,
|
||
|
|
"step": 478
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.479,
|
||
|
|
"grad_norm": 0.78125,
|
||
|
|
"learning_rate": 0.0001248689887164855,
|
||
|
|
"loss": 2.8572,
|
||
|
|
"step": 479
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.48,
|
||
|
|
"grad_norm": 0.78515625,
|
||
|
|
"learning_rate": 0.00012453073858788026,
|
||
|
|
"loss": 2.8889,
|
||
|
|
"step": 480
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.481,
|
||
|
|
"grad_norm": 0.7109375,
|
||
|
|
"learning_rate": 0.00012419218955996676,
|
||
|
|
"loss": 2.8363,
|
||
|
|
"step": 481
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.482,
|
||
|
|
"grad_norm": 0.7578125,
|
||
|
|
"learning_rate": 0.0001238533457578581,
|
||
|
|
"loss": 2.8624,
|
||
|
|
"step": 482
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.483,
|
||
|
|
"grad_norm": 0.7734375,
|
||
|
|
"learning_rate": 0.000123514211310259,
|
||
|
|
"loss": 2.8517,
|
||
|
|
"step": 483
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.484,
|
||
|
|
"grad_norm": 0.7421875,
|
||
|
|
"learning_rate": 0.00012317479034941573,
|
||
|
|
"loss": 2.8247,
|
||
|
|
"step": 484
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.485,
|
||
|
|
"grad_norm": 2.140625,
|
||
|
|
"learning_rate": 0.00012283508701106557,
|
||
|
|
"loss": 2.8114,
|
||
|
|
"step": 485
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.486,
|
||
|
|
"grad_norm": 0.8515625,
|
||
|
|
"learning_rate": 0.0001224951054343865,
|
||
|
|
"loss": 2.7308,
|
||
|
|
"step": 486
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.487,
|
||
|
|
"grad_norm": 0.76171875,
|
||
|
|
"learning_rate": 0.00012215484976194676,
|
||
|
|
"loss": 2.8445,
|
||
|
|
"step": 487
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.488,
|
||
|
|
"grad_norm": 0.8828125,
|
||
|
|
"learning_rate": 0.00012181432413965428,
|
||
|
|
"loss": 2.933,
|
||
|
|
"step": 488
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.489,
|
||
|
|
"grad_norm": 0.72265625,
|
||
|
|
"learning_rate": 0.00012147353271670634,
|
||
|
|
"loss": 2.8725,
|
||
|
|
"step": 489
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.49,
|
||
|
|
"grad_norm": 0.796875,
|
||
|
|
"learning_rate": 0.00012113247964553888,
|
||
|
|
"loss": 2.7981,
|
||
|
|
"step": 490
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.491,
|
||
|
|
"grad_norm": 0.74609375,
|
||
|
|
"learning_rate": 0.00012079116908177593,
|
||
|
|
"loss": 2.8803,
|
||
|
|
"step": 491
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.492,
|
||
|
|
"grad_norm": 0.73046875,
|
||
|
|
"learning_rate": 0.00012044960518417903,
|
||
|
|
"loss": 2.7637,
|
||
|
|
"step": 492
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.493,
|
||
|
|
"grad_norm": 0.75,
|
||
|
|
"learning_rate": 0.00012010779211459648,
|
||
|
|
"loss": 2.888,
|
||
|
|
"step": 493
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.494,
|
||
|
|
"grad_norm": 0.7265625,
|
||
|
|
"learning_rate": 0.00011976573403791262,
|
||
|
|
"loss": 2.815,
|
||
|
|
"step": 494
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.495,
|
||
|
|
"grad_norm": 0.77734375,
|
||
|
|
"learning_rate": 0.0001194234351219972,
|
||
|
|
"loss": 2.7684,
|
||
|
|
"step": 495
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.496,
|
||
|
|
"grad_norm": 0.6953125,
|
||
|
|
"learning_rate": 0.00011908089953765449,
|
||
|
|
"loss": 2.7292,
|
||
|
|
"step": 496
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.497,
|
||
|
|
"grad_norm": 0.69921875,
|
||
|
|
"learning_rate": 0.00011873813145857249,
|
||
|
|
"loss": 2.7949,
|
||
|
|
"step": 497
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.498,
|
||
|
|
"grad_norm": 0.6953125,
|
||
|
|
"learning_rate": 0.00011839513506127203,
|
||
|
|
"loss": 2.7681,
|
||
|
|
"step": 498
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.499,
|
||
|
|
"grad_norm": 0.69921875,
|
||
|
|
"learning_rate": 0.00011805191452505602,
|
||
|
|
"loss": 2.6686,
|
||
|
|
"step": 499
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.5,
|
||
|
|
"grad_norm": 1.2578125,
|
||
|
|
"learning_rate": 0.00011770847403195834,
|
||
|
|
"loss": 2.7635,
|
||
|
|
"step": 500
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.5,
|
||
|
|
"eval_loss": 2.7894091606140137,
|
||
|
|
"eval_runtime": 31.8024,
|
||
|
|
"eval_samples_per_second": 15.722,
|
||
|
|
"eval_steps_per_second": 2.641,
|
||
|
|
"step": 500
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.501,
|
||
|
|
"grad_norm": 0.7421875,
|
||
|
|
"learning_rate": 0.00011736481776669306,
|
||
|
|
"loss": 2.7591,
|
||
|
|
"step": 501
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.502,
|
||
|
|
"grad_norm": 0.73828125,
|
||
|
|
"learning_rate": 0.00011702094991660326,
|
||
|
|
"loss": 2.8107,
|
||
|
|
"step": 502
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.503,
|
||
|
|
"grad_norm": 0.734375,
|
||
|
|
"learning_rate": 0.00011667687467161024,
|
||
|
|
"loss": 2.8147,
|
||
|
|
"step": 503
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.504,
|
||
|
|
"grad_norm": 0.76171875,
|
||
|
|
"learning_rate": 0.00011633259622416224,
|
||
|
|
"loss": 2.8342,
|
||
|
|
"step": 504
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.505,
|
||
|
|
"grad_norm": 0.79296875,
|
||
|
|
"learning_rate": 0.0001159881187691835,
|
||
|
|
"loss": 2.8037,
|
||
|
|
"step": 505
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.506,
|
||
|
|
"grad_norm": 0.734375,
|
||
|
|
"learning_rate": 0.0001156434465040231,
|
||
|
|
"loss": 2.8258,
|
||
|
|
"step": 506
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.507,
|
||
|
|
"grad_norm": 0.84375,
|
||
|
|
"learning_rate": 0.00011529858362840382,
|
||
|
|
"loss": 2.8476,
|
||
|
|
"step": 507
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.508,
|
||
|
|
"grad_norm": 0.74609375,
|
||
|
|
"learning_rate": 0.00011495353434437098,
|
||
|
|
"loss": 2.7726,
|
||
|
|
"step": 508
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.509,
|
||
|
|
"grad_norm": 0.7890625,
|
||
|
|
"learning_rate": 0.00011460830285624118,
|
||
|
|
"loss": 2.7936,
|
||
|
|
"step": 509
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.51,
|
||
|
|
"grad_norm": 0.71875,
|
||
|
|
"learning_rate": 0.00011426289337055119,
|
||
|
|
"loss": 2.7585,
|
||
|
|
"step": 510
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.511,
|
||
|
|
"grad_norm": 0.703125,
|
||
|
|
"learning_rate": 0.00011391731009600654,
|
||
|
|
"loss": 2.7722,
|
||
|
|
"step": 511
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.512,
|
||
|
|
"grad_norm": 0.81640625,
|
||
|
|
"learning_rate": 0.00011357155724343045,
|
||
|
|
"loss": 2.8518,
|
||
|
|
"step": 512
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.513,
|
||
|
|
"grad_norm": 0.6484375,
|
||
|
|
"learning_rate": 0.00011322563902571226,
|
||
|
|
"loss": 2.8035,
|
||
|
|
"step": 513
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.514,
|
||
|
|
"grad_norm": 0.72265625,
|
||
|
|
"learning_rate": 0.0001128795596577563,
|
||
|
|
"loss": 2.8544,
|
||
|
|
"step": 514
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.515,
|
||
|
|
"grad_norm": 0.73828125,
|
||
|
|
"learning_rate": 0.00011253332335643043,
|
||
|
|
"loss": 2.8212,
|
||
|
|
"step": 515
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.516,
|
||
|
|
"grad_norm": 2.21875,
|
||
|
|
"learning_rate": 0.00011218693434051475,
|
||
|
|
"loss": 2.7369,
|
||
|
|
"step": 516
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.517,
|
||
|
|
"grad_norm": 0.8203125,
|
||
|
|
"learning_rate": 0.00011184039683065013,
|
||
|
|
"loss": 2.7366,
|
||
|
|
"step": 517
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.518,
|
||
|
|
"grad_norm": 0.73046875,
|
||
|
|
"learning_rate": 0.00011149371504928668,
|
||
|
|
"loss": 2.8155,
|
||
|
|
"step": 518
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.519,
|
||
|
|
"grad_norm": 0.77734375,
|
||
|
|
"learning_rate": 0.00011114689322063255,
|
||
|
|
"loss": 2.7464,
|
||
|
|
"step": 519
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.52,
|
||
|
|
"grad_norm": 0.8203125,
|
||
|
|
"learning_rate": 0.0001107999355706023,
|
||
|
|
"loss": 2.796,
|
||
|
|
"step": 520
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.521,
|
||
|
|
"grad_norm": 0.71875,
|
||
|
|
"learning_rate": 0.00011045284632676536,
|
||
|
|
"loss": 2.7958,
|
||
|
|
"step": 521
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.522,
|
||
|
|
"grad_norm": 0.703125,
|
||
|
|
"learning_rate": 0.00011010562971829463,
|
||
|
|
"loss": 2.7453,
|
||
|
|
"step": 522
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.523,
|
||
|
|
"grad_norm": 0.7421875,
|
||
|
|
"learning_rate": 0.00010975828997591495,
|
||
|
|
"loss": 2.7748,
|
||
|
|
"step": 523
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.524,
|
||
|
|
"grad_norm": 0.65625,
|
||
|
|
"learning_rate": 0.00010941083133185146,
|
||
|
|
"loss": 2.8185,
|
||
|
|
"step": 524
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.525,
|
||
|
|
"grad_norm": 0.7890625,
|
||
|
|
"learning_rate": 0.00010906325801977804,
|
||
|
|
"loss": 2.7946,
|
||
|
|
"step": 525
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.526,
|
||
|
|
"grad_norm": 0.71875,
|
||
|
|
"learning_rate": 0.00010871557427476583,
|
||
|
|
"loss": 2.8366,
|
||
|
|
"step": 526
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.527,
|
||
|
|
"grad_norm": 0.63671875,
|
||
|
|
"learning_rate": 0.00010836778433323158,
|
||
|
|
"loss": 2.81,
|
||
|
|
"step": 527
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.528,
|
||
|
|
"grad_norm": 0.7109375,
|
||
|
|
"learning_rate": 0.00010801989243288589,
|
||
|
|
"loss": 2.7869,
|
||
|
|
"step": 528
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.529,
|
||
|
|
"grad_norm": 0.6953125,
|
||
|
|
"learning_rate": 0.00010767190281268187,
|
||
|
|
"loss": 2.7461,
|
||
|
|
"step": 529
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.53,
|
||
|
|
"grad_norm": 0.6875,
|
||
|
|
"learning_rate": 0.00010732381971276318,
|
||
|
|
"loss": 2.7465,
|
||
|
|
"step": 530
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.531,
|
||
|
|
"grad_norm": 0.91796875,
|
||
|
|
"learning_rate": 0.00010697564737441252,
|
||
|
|
"loss": 2.7434,
|
||
|
|
"step": 531
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.532,
|
||
|
|
"grad_norm": 6.3125,
|
||
|
|
"learning_rate": 0.00010662739004000005,
|
||
|
|
"loss": 2.7999,
|
||
|
|
"step": 532
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.533,
|
||
|
|
"grad_norm": 0.6953125,
|
||
|
|
"learning_rate": 0.00010627905195293135,
|
||
|
|
"loss": 2.7405,
|
||
|
|
"step": 533
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.534,
|
||
|
|
"grad_norm": 0.69921875,
|
||
|
|
"learning_rate": 0.00010593063735759618,
|
||
|
|
"loss": 2.8709,
|
||
|
|
"step": 534
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.535,
|
||
|
|
"grad_norm": 0.66796875,
|
||
|
|
"learning_rate": 0.00010558215049931638,
|
||
|
|
"loss": 2.9645,
|
||
|
|
"step": 535
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.536,
|
||
|
|
"grad_norm": 0.7265625,
|
||
|
|
"learning_rate": 0.0001052335956242944,
|
||
|
|
"loss": 2.7778,
|
||
|
|
"step": 536
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.537,
|
||
|
|
"grad_norm": 0.66796875,
|
||
|
|
"learning_rate": 0.00010488497697956135,
|
||
|
|
"loss": 2.7532,
|
||
|
|
"step": 537
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.538,
|
||
|
|
"grad_norm": 0.8203125,
|
||
|
|
"learning_rate": 0.00010453629881292538,
|
||
|
|
"loss": 2.8057,
|
||
|
|
"step": 538
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.539,
|
||
|
|
"grad_norm": 0.70703125,
|
||
|
|
"learning_rate": 0.00010418756537291996,
|
||
|
|
"loss": 2.75,
|
||
|
|
"step": 539
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.54,
|
||
|
|
"grad_norm": 0.7109375,
|
||
|
|
"learning_rate": 0.00010383878090875201,
|
||
|
|
"loss": 2.8593,
|
||
|
|
"step": 540
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.541,
|
||
|
|
"grad_norm": 0.6640625,
|
||
|
|
"learning_rate": 0.00010348994967025012,
|
||
|
|
"loss": 2.7264,
|
||
|
|
"step": 541
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.542,
|
||
|
|
"grad_norm": 0.703125,
|
||
|
|
"learning_rate": 0.00010314107590781284,
|
||
|
|
"loss": 2.8229,
|
||
|
|
"step": 542
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.543,
|
||
|
|
"grad_norm": 0.68359375,
|
||
|
|
"learning_rate": 0.0001027921638723569,
|
||
|
|
"loss": 2.7506,
|
||
|
|
"step": 543
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.544,
|
||
|
|
"grad_norm": 0.65625,
|
||
|
|
"learning_rate": 0.00010244321781526533,
|
||
|
|
"loss": 2.7689,
|
||
|
|
"step": 544
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.545,
|
||
|
|
"grad_norm": 0.734375,
|
||
|
|
"learning_rate": 0.0001020942419883357,
|
||
|
|
"loss": 2.766,
|
||
|
|
"step": 545
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.546,
|
||
|
|
"grad_norm": 0.69140625,
|
||
|
|
"learning_rate": 0.00010174524064372837,
|
||
|
|
"loss": 2.8915,
|
||
|
|
"step": 546
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.547,
|
||
|
|
"grad_norm": 2.625,
|
||
|
|
"learning_rate": 0.00010139621803391455,
|
||
|
|
"loss": 2.7575,
|
||
|
|
"step": 547
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.548,
|
||
|
|
"grad_norm": 0.7578125,
|
||
|
|
"learning_rate": 0.00010104717841162458,
|
||
|
|
"loss": 2.7613,
|
||
|
|
"step": 548
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.549,
|
||
|
|
"grad_norm": 0.6875,
|
||
|
|
"learning_rate": 0.00010069812602979615,
|
||
|
|
"loss": 2.8354,
|
||
|
|
"step": 549
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.55,
|
||
|
|
"grad_norm": 0.640625,
|
||
|
|
"learning_rate": 0.00010034906514152238,
|
||
|
|
"loss": 2.815,
|
||
|
|
"step": 550
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.551,
|
||
|
|
"grad_norm": 0.78515625,
|
||
|
|
"learning_rate": 0.0001,
|
||
|
|
"loss": 2.9268,
|
||
|
|
"step": 551
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.552,
|
||
|
|
"grad_norm": 0.7109375,
|
||
|
|
"learning_rate": 9.965093485847767e-05,
|
||
|
|
"loss": 2.7682,
|
||
|
|
"step": 552
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.553,
|
||
|
|
"grad_norm": 0.6484375,
|
||
|
|
"learning_rate": 9.930187397020386e-05,
|
||
|
|
"loss": 2.8294,
|
||
|
|
"step": 553
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.554,
|
||
|
|
"grad_norm": 0.734375,
|
||
|
|
"learning_rate": 9.895282158837545e-05,
|
||
|
|
"loss": 2.7828,
|
||
|
|
"step": 554
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.555,
|
||
|
|
"grad_norm": 0.6953125,
|
||
|
|
"learning_rate": 9.860378196608549e-05,
|
||
|
|
"loss": 2.7539,
|
||
|
|
"step": 555
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.556,
|
||
|
|
"grad_norm": 0.6953125,
|
||
|
|
"learning_rate": 9.825475935627165e-05,
|
||
|
|
"loss": 2.7614,
|
||
|
|
"step": 556
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.557,
|
||
|
|
"grad_norm": 0.73828125,
|
||
|
|
"learning_rate": 9.790575801166432e-05,
|
||
|
|
"loss": 2.7728,
|
||
|
|
"step": 557
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.558,
|
||
|
|
"grad_norm": 0.7265625,
|
||
|
|
"learning_rate": 9.755678218473469e-05,
|
||
|
|
"loss": 2.7955,
|
||
|
|
"step": 558
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.559,
|
||
|
|
"grad_norm": 0.6796875,
|
||
|
|
"learning_rate": 9.720783612764314e-05,
|
||
|
|
"loss": 2.8389,
|
||
|
|
"step": 559
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.56,
|
||
|
|
"grad_norm": 0.70703125,
|
||
|
|
"learning_rate": 9.685892409218717e-05,
|
||
|
|
"loss": 2.7221,
|
||
|
|
"step": 560
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.561,
|
||
|
|
"grad_norm": 0.69140625,
|
||
|
|
"learning_rate": 9.651005032974994e-05,
|
||
|
|
"loss": 2.7392,
|
||
|
|
"step": 561
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.562,
|
||
|
|
"grad_norm": 0.65625,
|
||
|
|
"learning_rate": 9.616121909124801e-05,
|
||
|
|
"loss": 2.7522,
|
||
|
|
"step": 562
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.563,
|
||
|
|
"grad_norm": 4.125,
|
||
|
|
"learning_rate": 9.581243462708006e-05,
|
||
|
|
"loss": 2.7441,
|
||
|
|
"step": 563
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.564,
|
||
|
|
"grad_norm": 0.81640625,
|
||
|
|
"learning_rate": 9.546370118707463e-05,
|
||
|
|
"loss": 2.7218,
|
||
|
|
"step": 564
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.565,
|
||
|
|
"grad_norm": 0.73046875,
|
||
|
|
"learning_rate": 9.511502302043868e-05,
|
||
|
|
"loss": 2.7805,
|
||
|
|
"step": 565
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.566,
|
||
|
|
"grad_norm": 0.67578125,
|
||
|
|
"learning_rate": 9.476640437570562e-05,
|
||
|
|
"loss": 2.7791,
|
||
|
|
"step": 566
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.567,
|
||
|
|
"grad_norm": 0.69921875,
|
||
|
|
"learning_rate": 9.441784950068362e-05,
|
||
|
|
"loss": 2.8229,
|
||
|
|
"step": 567
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.568,
|
||
|
|
"grad_norm": 0.71875,
|
||
|
|
"learning_rate": 9.406936264240386e-05,
|
||
|
|
"loss": 2.8902,
|
||
|
|
"step": 568
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.569,
|
||
|
|
"grad_norm": 0.71484375,
|
||
|
|
"learning_rate": 9.372094804706867e-05,
|
||
|
|
"loss": 2.8009,
|
||
|
|
"step": 569
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.57,
|
||
|
|
"grad_norm": 0.62890625,
|
||
|
|
"learning_rate": 9.337260996000002e-05,
|
||
|
|
"loss": 2.7867,
|
||
|
|
"step": 570
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.571,
|
||
|
|
"grad_norm": 0.71875,
|
||
|
|
"learning_rate": 9.302435262558747e-05,
|
||
|
|
"loss": 2.7513,
|
||
|
|
"step": 571
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.572,
|
||
|
|
"grad_norm": 0.6875,
|
||
|
|
"learning_rate": 9.267618028723686e-05,
|
||
|
|
"loss": 2.8066,
|
||
|
|
"step": 572
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.573,
|
||
|
|
"grad_norm": 0.6484375,
|
||
|
|
"learning_rate": 9.232809718731814e-05,
|
||
|
|
"loss": 2.818,
|
||
|
|
"step": 573
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.574,
|
||
|
|
"grad_norm": 0.69921875,
|
||
|
|
"learning_rate": 9.198010756711412e-05,
|
||
|
|
"loss": 2.6698,
|
||
|
|
"step": 574
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.575,
|
||
|
|
"grad_norm": 0.7109375,
|
||
|
|
"learning_rate": 9.163221566676847e-05,
|
||
|
|
"loss": 2.7901,
|
||
|
|
"step": 575
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.576,
|
||
|
|
"grad_norm": 2.0625,
|
||
|
|
"learning_rate": 9.128442572523417e-05,
|
||
|
|
"loss": 2.8648,
|
||
|
|
"step": 576
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.577,
|
||
|
|
"grad_norm": 0.78515625,
|
||
|
|
"learning_rate": 9.093674198022201e-05,
|
||
|
|
"loss": 2.8698,
|
||
|
|
"step": 577
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.578,
|
||
|
|
"grad_norm": 0.66796875,
|
||
|
|
"learning_rate": 9.058916866814858e-05,
|
||
|
|
"loss": 2.8594,
|
||
|
|
"step": 578
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.579,
|
||
|
|
"grad_norm": 3.65625,
|
||
|
|
"learning_rate": 9.024171002408506e-05,
|
||
|
|
"loss": 2.788,
|
||
|
|
"step": 579
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.58,
|
||
|
|
"grad_norm": 0.8125,
|
||
|
|
"learning_rate": 8.989437028170537e-05,
|
||
|
|
"loss": 2.7126,
|
||
|
|
"step": 580
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.581,
|
||
|
|
"grad_norm": 0.734375,
|
||
|
|
"learning_rate": 8.954715367323468e-05,
|
||
|
|
"loss": 2.7795,
|
||
|
|
"step": 581
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.582,
|
||
|
|
"grad_norm": 0.72265625,
|
||
|
|
"learning_rate": 8.920006442939772e-05,
|
||
|
|
"loss": 2.7243,
|
||
|
|
"step": 582
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.583,
|
||
|
|
"grad_norm": 0.828125,
|
||
|
|
"learning_rate": 8.885310677936746e-05,
|
||
|
|
"loss": 2.8055,
|
||
|
|
"step": 583
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.584,
|
||
|
|
"grad_norm": 0.66796875,
|
||
|
|
"learning_rate": 8.850628495071336e-05,
|
||
|
|
"loss": 2.8449,
|
||
|
|
"step": 584
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.585,
|
||
|
|
"grad_norm": 0.83203125,
|
||
|
|
"learning_rate": 8.81596031693499e-05,
|
||
|
|
"loss": 2.7153,
|
||
|
|
"step": 585
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.586,
|
||
|
|
"grad_norm": 0.7890625,
|
||
|
|
"learning_rate": 8.781306565948528e-05,
|
||
|
|
"loss": 2.8445,
|
||
|
|
"step": 586
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.587,
|
||
|
|
"grad_norm": 0.60546875,
|
||
|
|
"learning_rate": 8.746667664356956e-05,
|
||
|
|
"loss": 2.7783,
|
||
|
|
"step": 587
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.588,
|
||
|
|
"grad_norm": 0.6953125,
|
||
|
|
"learning_rate": 8.712044034224374e-05,
|
||
|
|
"loss": 2.8369,
|
||
|
|
"step": 588
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.589,
|
||
|
|
"grad_norm": 0.92578125,
|
||
|
|
"learning_rate": 8.677436097428775e-05,
|
||
|
|
"loss": 2.7912,
|
||
|
|
"step": 589
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.59,
|
||
|
|
"grad_norm": 0.6484375,
|
||
|
|
"learning_rate": 8.642844275656957e-05,
|
||
|
|
"loss": 2.8142,
|
||
|
|
"step": 590
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.591,
|
||
|
|
"grad_norm": 0.72265625,
|
||
|
|
"learning_rate": 8.608268990399349e-05,
|
||
|
|
"loss": 2.6924,
|
||
|
|
"step": 591
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.592,
|
||
|
|
"grad_norm": 0.734375,
|
||
|
|
"learning_rate": 8.573710662944885e-05,
|
||
|
|
"loss": 2.7975,
|
||
|
|
"step": 592
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.593,
|
||
|
|
"grad_norm": 0.69140625,
|
||
|
|
"learning_rate": 8.539169714375885e-05,
|
||
|
|
"loss": 2.8273,
|
||
|
|
"step": 593
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.594,
|
||
|
|
"grad_norm": 3.609375,
|
||
|
|
"learning_rate": 8.504646565562906e-05,
|
||
|
|
"loss": 2.7106,
|
||
|
|
"step": 594
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.595,
|
||
|
|
"grad_norm": 0.8359375,
|
||
|
|
"learning_rate": 8.47014163715962e-05,
|
||
|
|
"loss": 2.773,
|
||
|
|
"step": 595
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.596,
|
||
|
|
"grad_norm": 0.8203125,
|
||
|
|
"learning_rate": 8.435655349597689e-05,
|
||
|
|
"loss": 2.759,
|
||
|
|
"step": 596
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.597,
|
||
|
|
"grad_norm": 0.7265625,
|
||
|
|
"learning_rate": 8.401188123081653e-05,
|
||
|
|
"loss": 2.7223,
|
||
|
|
"step": 597
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.598,
|
||
|
|
"grad_norm": 0.73828125,
|
||
|
|
"learning_rate": 8.366740377583781e-05,
|
||
|
|
"loss": 2.7163,
|
||
|
|
"step": 598
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.599,
|
||
|
|
"grad_norm": 0.76953125,
|
||
|
|
"learning_rate": 8.332312532838978e-05,
|
||
|
|
"loss": 2.7544,
|
||
|
|
"step": 599
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.6,
|
||
|
|
"grad_norm": 0.71484375,
|
||
|
|
"learning_rate": 8.297905008339677e-05,
|
||
|
|
"loss": 2.7424,
|
||
|
|
"step": 600
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.601,
|
||
|
|
"grad_norm": 0.70703125,
|
||
|
|
"learning_rate": 8.263518223330697e-05,
|
||
|
|
"loss": 2.8959,
|
||
|
|
"step": 601
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.602,
|
||
|
|
"grad_norm": 0.6640625,
|
||
|
|
"learning_rate": 8.229152596804168e-05,
|
||
|
|
"loss": 2.665,
|
||
|
|
"step": 602
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.603,
|
||
|
|
"grad_norm": 0.6640625,
|
||
|
|
"learning_rate": 8.194808547494401e-05,
|
||
|
|
"loss": 2.8718,
|
||
|
|
"step": 603
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.604,
|
||
|
|
"grad_norm": 0.66796875,
|
||
|
|
"learning_rate": 8.160486493872798e-05,
|
||
|
|
"loss": 2.7585,
|
||
|
|
"step": 604
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.605,
|
||
|
|
"grad_norm": 0.65625,
|
||
|
|
"learning_rate": 8.126186854142752e-05,
|
||
|
|
"loss": 2.7911,
|
||
|
|
"step": 605
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.606,
|
||
|
|
"grad_norm": 0.92578125,
|
||
|
|
"learning_rate": 8.091910046234552e-05,
|
||
|
|
"loss": 2.6548,
|
||
|
|
"step": 606
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.607,
|
||
|
|
"grad_norm": 0.625,
|
||
|
|
"learning_rate": 8.057656487800282e-05,
|
||
|
|
"loss": 2.7227,
|
||
|
|
"step": 607
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.608,
|
||
|
|
"grad_norm": 0.66796875,
|
||
|
|
"learning_rate": 8.023426596208739e-05,
|
||
|
|
"loss": 2.7485,
|
||
|
|
"step": 608
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.609,
|
||
|
|
"grad_norm": 0.65625,
|
||
|
|
"learning_rate": 7.989220788540355e-05,
|
||
|
|
"loss": 2.7348,
|
||
|
|
"step": 609
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.61,
|
||
|
|
"grad_norm": 1.546875,
|
||
|
|
"learning_rate": 7.955039481582097e-05,
|
||
|
|
"loss": 2.7158,
|
||
|
|
"step": 610
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.611,
|
||
|
|
"grad_norm": 0.69921875,
|
||
|
|
"learning_rate": 7.920883091822408e-05,
|
||
|
|
"loss": 2.6516,
|
||
|
|
"step": 611
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.612,
|
||
|
|
"grad_norm": 0.6484375,
|
||
|
|
"learning_rate": 7.886752035446114e-05,
|
||
|
|
"loss": 2.8051,
|
||
|
|
"step": 612
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.613,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 7.852646728329368e-05,
|
||
|
|
"loss": 2.7428,
|
||
|
|
"step": 613
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.614,
|
||
|
|
"grad_norm": 0.70703125,
|
||
|
|
"learning_rate": 7.818567586034577e-05,
|
||
|
|
"loss": 2.8212,
|
||
|
|
"step": 614
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.615,
|
||
|
|
"grad_norm": 0.66015625,
|
||
|
|
"learning_rate": 7.784515023805328e-05,
|
||
|
|
"loss": 2.7348,
|
||
|
|
"step": 615
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.616,
|
||
|
|
"grad_norm": 0.69921875,
|
||
|
|
"learning_rate": 7.750489456561352e-05,
|
||
|
|
"loss": 2.7394,
|
||
|
|
"step": 616
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.617,
|
||
|
|
"grad_norm": 0.67578125,
|
||
|
|
"learning_rate": 7.716491298893442e-05,
|
||
|
|
"loss": 2.6866,
|
||
|
|
"step": 617
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.618,
|
||
|
|
"grad_norm": 0.67578125,
|
||
|
|
"learning_rate": 7.682520965058428e-05,
|
||
|
|
"loss": 2.6922,
|
||
|
|
"step": 618
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.619,
|
||
|
|
"grad_norm": 0.64453125,
|
||
|
|
"learning_rate": 7.6485788689741e-05,
|
||
|
|
"loss": 2.6569,
|
||
|
|
"step": 619
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.62,
|
||
|
|
"grad_norm": 0.640625,
|
||
|
|
"learning_rate": 7.614665424214193e-05,
|
||
|
|
"loss": 2.752,
|
||
|
|
"step": 620
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.621,
|
||
|
|
"grad_norm": 0.69921875,
|
||
|
|
"learning_rate": 7.580781044003324e-05,
|
||
|
|
"loss": 2.7281,
|
||
|
|
"step": 621
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.622,
|
||
|
|
"grad_norm": 1.171875,
|
||
|
|
"learning_rate": 7.546926141211974e-05,
|
||
|
|
"loss": 2.8616,
|
||
|
|
"step": 622
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.623,
|
||
|
|
"grad_norm": 0.6796875,
|
||
|
|
"learning_rate": 7.513101128351454e-05,
|
||
|
|
"loss": 2.8221,
|
||
|
|
"step": 623
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.624,
|
||
|
|
"grad_norm": 0.640625,
|
||
|
|
"learning_rate": 7.479306417568864e-05,
|
||
|
|
"loss": 2.8145,
|
||
|
|
"step": 624
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.625,
|
||
|
|
"grad_norm": 1.8359375,
|
||
|
|
"learning_rate": 7.445542420642097e-05,
|
||
|
|
"loss": 2.6816,
|
||
|
|
"step": 625
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.626,
|
||
|
|
"grad_norm": 0.640625,
|
||
|
|
"learning_rate": 7.411809548974792e-05,
|
||
|
|
"loss": 2.6947,
|
||
|
|
"step": 626
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.627,
|
||
|
|
"grad_norm": 0.66796875,
|
||
|
|
"learning_rate": 7.378108213591355e-05,
|
||
|
|
"loss": 2.7485,
|
||
|
|
"step": 627
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.628,
|
||
|
|
"grad_norm": 0.66015625,
|
||
|
|
"learning_rate": 7.344438825131911e-05,
|
||
|
|
"loss": 2.7634,
|
||
|
|
"step": 628
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.629,
|
||
|
|
"grad_norm": 0.60546875,
|
||
|
|
"learning_rate": 7.310801793847344e-05,
|
||
|
|
"loss": 2.7374,
|
||
|
|
"step": 629
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.63,
|
||
|
|
"grad_norm": 0.65625,
|
||
|
|
"learning_rate": 7.277197529594257e-05,
|
||
|
|
"loss": 2.6431,
|
||
|
|
"step": 630
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.631,
|
||
|
|
"grad_norm": 0.66796875,
|
||
|
|
"learning_rate": 7.243626441830009e-05,
|
||
|
|
"loss": 2.6069,
|
||
|
|
"step": 631
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.632,
|
||
|
|
"grad_norm": 0.6328125,
|
||
|
|
"learning_rate": 7.210088939607708e-05,
|
||
|
|
"loss": 2.7996,
|
||
|
|
"step": 632
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.633,
|
||
|
|
"grad_norm": 1.046875,
|
||
|
|
"learning_rate": 7.176585431571235e-05,
|
||
|
|
"loss": 2.7437,
|
||
|
|
"step": 633
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.634,
|
||
|
|
"grad_norm": 0.62890625,
|
||
|
|
"learning_rate": 7.143116325950265e-05,
|
||
|
|
"loss": 2.7624,
|
||
|
|
"step": 634
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.635,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 7.109682030555283e-05,
|
||
|
|
"loss": 2.715,
|
||
|
|
"step": 635
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.636,
|
||
|
|
"grad_norm": 0.70703125,
|
||
|
|
"learning_rate": 7.076282952772633e-05,
|
||
|
|
"loss": 2.7208,
|
||
|
|
"step": 636
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.637,
|
||
|
|
"grad_norm": 0.6328125,
|
||
|
|
"learning_rate": 7.042919499559537e-05,
|
||
|
|
"loss": 2.7576,
|
||
|
|
"step": 637
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.638,
|
||
|
|
"grad_norm": 0.58203125,
|
||
|
|
"learning_rate": 7.009592077439134e-05,
|
||
|
|
"loss": 2.8255,
|
||
|
|
"step": 638
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.639,
|
||
|
|
"grad_norm": 0.640625,
|
||
|
|
"learning_rate": 6.976301092495556e-05,
|
||
|
|
"loss": 2.7823,
|
||
|
|
"step": 639
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.64,
|
||
|
|
"grad_norm": 0.64453125,
|
||
|
|
"learning_rate": 6.943046950368944e-05,
|
||
|
|
"loss": 2.6574,
|
||
|
|
"step": 640
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.641,
|
||
|
|
"grad_norm": 1.2578125,
|
||
|
|
"learning_rate": 6.909830056250527e-05,
|
||
|
|
"loss": 2.6497,
|
||
|
|
"step": 641
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.642,
|
||
|
|
"grad_norm": 0.6171875,
|
||
|
|
"learning_rate": 6.876650814877674e-05,
|
||
|
|
"loss": 2.7926,
|
||
|
|
"step": 642
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.643,
|
||
|
|
"grad_norm": 0.63671875,
|
||
|
|
"learning_rate": 6.843509630528977e-05,
|
||
|
|
"loss": 2.6997,
|
||
|
|
"step": 643
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.644,
|
||
|
|
"grad_norm": 0.61328125,
|
||
|
|
"learning_rate": 6.8104069070193e-05,
|
||
|
|
"loss": 2.7251,
|
||
|
|
"step": 644
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.645,
|
||
|
|
"grad_norm": 0.62109375,
|
||
|
|
"learning_rate": 6.77734304769489e-05,
|
||
|
|
"loss": 2.6186,
|
||
|
|
"step": 645
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.646,
|
||
|
|
"grad_norm": 0.6171875,
|
||
|
|
"learning_rate": 6.744318455428436e-05,
|
||
|
|
"loss": 2.7101,
|
||
|
|
"step": 646
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.647,
|
||
|
|
"grad_norm": 0.66015625,
|
||
|
|
"learning_rate": 6.711333532614168e-05,
|
||
|
|
"loss": 2.6802,
|
||
|
|
"step": 647
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.648,
|
||
|
|
"grad_norm": 0.671875,
|
||
|
|
"learning_rate": 6.67838868116297e-05,
|
||
|
|
"loss": 2.7727,
|
||
|
|
"step": 648
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.649,
|
||
|
|
"grad_norm": 0.640625,
|
||
|
|
"learning_rate": 6.64548430249745e-05,
|
||
|
|
"loss": 2.8216,
|
||
|
|
"step": 649
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.65,
|
||
|
|
"grad_norm": 0.6640625,
|
||
|
|
"learning_rate": 6.612620797547087e-05,
|
||
|
|
"loss": 2.8565,
|
||
|
|
"step": 650
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.651,
|
||
|
|
"grad_norm": 0.671875,
|
||
|
|
"learning_rate": 6.579798566743314e-05,
|
||
|
|
"loss": 2.795,
|
||
|
|
"step": 651
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.652,
|
||
|
|
"grad_norm": 0.66015625,
|
||
|
|
"learning_rate": 6.547018010014654e-05,
|
||
|
|
"loss": 2.7028,
|
||
|
|
"step": 652
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.653,
|
||
|
|
"grad_norm": 0.625,
|
||
|
|
"learning_rate": 6.51427952678185e-05,
|
||
|
|
"loss": 2.7782,
|
||
|
|
"step": 653
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.654,
|
||
|
|
"grad_norm": 0.60546875,
|
||
|
|
"learning_rate": 6.481583515952983e-05,
|
||
|
|
"loss": 2.7242,
|
||
|
|
"step": 654
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.655,
|
||
|
|
"grad_norm": 0.6640625,
|
||
|
|
"learning_rate": 6.448930375918631e-05,
|
||
|
|
"loss": 2.7598,
|
||
|
|
"step": 655
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.656,
|
||
|
|
"grad_norm": 0.64453125,
|
||
|
|
"learning_rate": 6.416320504546997e-05,
|
||
|
|
"loss": 2.6908,
|
||
|
|
"step": 656
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.657,
|
||
|
|
"grad_norm": 1.265625,
|
||
|
|
"learning_rate": 6.383754299179079e-05,
|
||
|
|
"loss": 2.7308,
|
||
|
|
"step": 657
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.658,
|
||
|
|
"grad_norm": 0.66796875,
|
||
|
|
"learning_rate": 6.351232156623803e-05,
|
||
|
|
"loss": 2.7679,
|
||
|
|
"step": 658
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.659,
|
||
|
|
"grad_norm": 0.66015625,
|
||
|
|
"learning_rate": 6.318754473153221e-05,
|
||
|
|
"loss": 2.8747,
|
||
|
|
"step": 659
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.66,
|
||
|
|
"grad_norm": 0.62890625,
|
||
|
|
"learning_rate": 6.286321644497655e-05,
|
||
|
|
"loss": 2.7716,
|
||
|
|
"step": 660
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.661,
|
||
|
|
"grad_norm": 0.62890625,
|
||
|
|
"learning_rate": 6.25393406584088e-05,
|
||
|
|
"loss": 2.6582,
|
||
|
|
"step": 661
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.662,
|
||
|
|
"grad_norm": 0.60546875,
|
||
|
|
"learning_rate": 6.22159213181533e-05,
|
||
|
|
"loss": 2.7907,
|
||
|
|
"step": 662
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.663,
|
||
|
|
"grad_norm": 0.67578125,
|
||
|
|
"learning_rate": 6.18929623649726e-05,
|
||
|
|
"loss": 2.6852,
|
||
|
|
"step": 663
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.664,
|
||
|
|
"grad_norm": 0.61328125,
|
||
|
|
"learning_rate": 6.157046773401964e-05,
|
||
|
|
"loss": 2.7763,
|
||
|
|
"step": 664
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.665,
|
||
|
|
"grad_norm": 0.640625,
|
||
|
|
"learning_rate": 6.12484413547897e-05,
|
||
|
|
"loss": 2.713,
|
||
|
|
"step": 665
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.666,
|
||
|
|
"grad_norm": 0.75390625,
|
||
|
|
"learning_rate": 6.092688715107264e-05,
|
||
|
|
"loss": 2.8589,
|
||
|
|
"step": 666
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.667,
|
||
|
|
"grad_norm": 0.67578125,
|
||
|
|
"learning_rate": 6.0605809040904894e-05,
|
||
|
|
"loss": 2.813,
|
||
|
|
"step": 667
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.668,
|
||
|
|
"grad_norm": 0.68359375,
|
||
|
|
"learning_rate": 6.0285210936521955e-05,
|
||
|
|
"loss": 2.7389,
|
||
|
|
"step": 668
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.669,
|
||
|
|
"grad_norm": 0.625,
|
||
|
|
"learning_rate": 5.9965096744310526e-05,
|
||
|
|
"loss": 2.7049,
|
||
|
|
"step": 669
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.67,
|
||
|
|
"grad_norm": 0.6875,
|
||
|
|
"learning_rate": 5.964547036476099e-05,
|
||
|
|
"loss": 2.7361,
|
||
|
|
"step": 670
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.671,
|
||
|
|
"grad_norm": 0.61328125,
|
||
|
|
"learning_rate": 5.9326335692419995e-05,
|
||
|
|
"loss": 2.7189,
|
||
|
|
"step": 671
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.672,
|
||
|
|
"grad_norm": 1.5078125,
|
||
|
|
"learning_rate": 5.900769661584272e-05,
|
||
|
|
"loss": 2.7206,
|
||
|
|
"step": 672
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.673,
|
||
|
|
"grad_norm": 0.69921875,
|
||
|
|
"learning_rate": 5.868955701754584e-05,
|
||
|
|
"loss": 2.7581,
|
||
|
|
"step": 673
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.674,
|
||
|
|
"grad_norm": 0.63671875,
|
||
|
|
"learning_rate": 5.83719207739599e-05,
|
||
|
|
"loss": 2.5903,
|
||
|
|
"step": 674
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.675,
|
||
|
|
"grad_norm": 0.65234375,
|
||
|
|
"learning_rate": 5.805479175538229e-05,
|
||
|
|
"loss": 2.7786,
|
||
|
|
"step": 675
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.676,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 5.773817382593008e-05,
|
||
|
|
"loss": 2.7335,
|
||
|
|
"step": 676
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.677,
|
||
|
|
"grad_norm": 0.6171875,
|
||
|
|
"learning_rate": 5.7422070843492734e-05,
|
||
|
|
"loss": 2.6639,
|
||
|
|
"step": 677
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.678,
|
||
|
|
"grad_norm": 0.66015625,
|
||
|
|
"learning_rate": 5.710648665968543e-05,
|
||
|
|
"loss": 2.6919,
|
||
|
|
"step": 678
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.679,
|
||
|
|
"grad_norm": 0.6328125,
|
||
|
|
"learning_rate": 5.679142511980175e-05,
|
||
|
|
"loss": 2.7755,
|
||
|
|
"step": 679
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.68,
|
||
|
|
"grad_norm": 0.6171875,
|
||
|
|
"learning_rate": 5.647689006276726e-05,
|
||
|
|
"loss": 2.8639,
|
||
|
|
"step": 680
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.681,
|
||
|
|
"grad_norm": 0.6953125,
|
||
|
|
"learning_rate": 5.616288532109225e-05,
|
||
|
|
"loss": 2.9127,
|
||
|
|
"step": 681
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.682,
|
||
|
|
"grad_norm": 0.73046875,
|
||
|
|
"learning_rate": 5.584941472082549e-05,
|
||
|
|
"loss": 2.8527,
|
||
|
|
"step": 682
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.683,
|
||
|
|
"grad_norm": 0.6875,
|
||
|
|
"learning_rate": 5.553648208150728e-05,
|
||
|
|
"loss": 2.7896,
|
||
|
|
"step": 683
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.684,
|
||
|
|
"grad_norm": 0.6171875,
|
||
|
|
"learning_rate": 5.522409121612304e-05,
|
||
|
|
"loss": 2.7857,
|
||
|
|
"step": 684
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.685,
|
||
|
|
"grad_norm": 0.65625,
|
||
|
|
"learning_rate": 5.491224593105695e-05,
|
||
|
|
"loss": 2.7679,
|
||
|
|
"step": 685
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.686,
|
||
|
|
"grad_norm": 0.6640625,
|
||
|
|
"learning_rate": 5.4600950026045326e-05,
|
||
|
|
"loss": 2.7092,
|
||
|
|
"step": 686
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.687,
|
||
|
|
"grad_norm": 0.64453125,
|
||
|
|
"learning_rate": 5.4290207294130615e-05,
|
||
|
|
"loss": 2.7738,
|
||
|
|
"step": 687
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.688,
|
||
|
|
"grad_norm": 0.99609375,
|
||
|
|
"learning_rate": 5.398002152161484e-05,
|
||
|
|
"loss": 2.7504,
|
||
|
|
"step": 688
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.689,
|
||
|
|
"grad_norm": 0.609375,
|
||
|
|
"learning_rate": 5.3670396488013854e-05,
|
||
|
|
"loss": 2.783,
|
||
|
|
"step": 689
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.69,
|
||
|
|
"grad_norm": 0.640625,
|
||
|
|
"learning_rate": 5.33613359660109e-05,
|
||
|
|
"loss": 2.7702,
|
||
|
|
"step": 690
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.691,
|
||
|
|
"grad_norm": 0.66796875,
|
||
|
|
"learning_rate": 5.305284372141095e-05,
|
||
|
|
"loss": 2.7703,
|
||
|
|
"step": 691
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.692,
|
||
|
|
"grad_norm": 0.67578125,
|
||
|
|
"learning_rate": 5.274492351309461e-05,
|
||
|
|
"loss": 2.8481,
|
||
|
|
"step": 692
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.693,
|
||
|
|
"grad_norm": 0.640625,
|
||
|
|
"learning_rate": 5.243757909297247e-05,
|
||
|
|
"loss": 2.8131,
|
||
|
|
"step": 693
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.694,
|
||
|
|
"grad_norm": 0.62109375,
|
||
|
|
"learning_rate": 5.213081420593933e-05,
|
||
|
|
"loss": 2.8698,
|
||
|
|
"step": 694
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.695,
|
||
|
|
"grad_norm": 0.68359375,
|
||
|
|
"learning_rate": 5.182463258982846e-05,
|
||
|
|
"loss": 2.8107,
|
||
|
|
"step": 695
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.696,
|
||
|
|
"grad_norm": 0.6796875,
|
||
|
|
"learning_rate": 5.15190379753663e-05,
|
||
|
|
"loss": 2.8036,
|
||
|
|
"step": 696
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.697,
|
||
|
|
"grad_norm": 0.64453125,
|
||
|
|
"learning_rate": 5.121403408612672e-05,
|
||
|
|
"loss": 2.8471,
|
||
|
|
"step": 697
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.698,
|
||
|
|
"grad_norm": 0.609375,
|
||
|
|
"learning_rate": 5.090962463848592e-05,
|
||
|
|
"loss": 2.8396,
|
||
|
|
"step": 698
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.699,
|
||
|
|
"grad_norm": 0.62890625,
|
||
|
|
"learning_rate": 5.0605813341576924e-05,
|
||
|
|
"loss": 2.8636,
|
||
|
|
"step": 699
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.7,
|
||
|
|
"grad_norm": 0.70703125,
|
||
|
|
"learning_rate": 5.0302603897244474e-05,
|
||
|
|
"loss": 2.9405,
|
||
|
|
"step": 700
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.701,
|
||
|
|
"grad_norm": 0.62890625,
|
||
|
|
"learning_rate": 5.000000000000002e-05,
|
||
|
|
"loss": 2.7732,
|
||
|
|
"step": 701
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.702,
|
||
|
|
"grad_norm": 0.671875,
|
||
|
|
"learning_rate": 4.969800533697649e-05,
|
||
|
|
"loss": 2.9202,
|
||
|
|
"step": 702
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.703,
|
||
|
|
"grad_norm": 0.65234375,
|
||
|
|
"learning_rate": 4.939662358788364e-05,
|
||
|
|
"loss": 2.8386,
|
||
|
|
"step": 703
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.704,
|
||
|
|
"grad_norm": 0.82421875,
|
||
|
|
"learning_rate": 4.909585842496287e-05,
|
||
|
|
"loss": 2.8088,
|
||
|
|
"step": 704
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.705,
|
||
|
|
"grad_norm": 0.66796875,
|
||
|
|
"learning_rate": 4.8795713512942865e-05,
|
||
|
|
"loss": 2.6467,
|
||
|
|
"step": 705
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.706,
|
||
|
|
"grad_norm": 0.65234375,
|
||
|
|
"learning_rate": 4.8496192508994576e-05,
|
||
|
|
"loss": 2.7861,
|
||
|
|
"step": 706
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.707,
|
||
|
|
"grad_norm": 0.52734375,
|
||
|
|
"learning_rate": 4.8197299062686995e-05,
|
||
|
|
"loss": 2.7869,
|
||
|
|
"step": 707
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.708,
|
||
|
|
"grad_norm": 0.6015625,
|
||
|
|
"learning_rate": 4.78990368159424e-05,
|
||
|
|
"loss": 2.6457,
|
||
|
|
"step": 708
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.709,
|
||
|
|
"grad_norm": 0.57421875,
|
||
|
|
"learning_rate": 4.7601409402992106e-05,
|
||
|
|
"loss": 2.6031,
|
||
|
|
"step": 709
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.71,
|
||
|
|
"grad_norm": 0.58203125,
|
||
|
|
"learning_rate": 4.7304420450332244e-05,
|
||
|
|
"loss": 2.8342,
|
||
|
|
"step": 710
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.711,
|
||
|
|
"grad_norm": 0.56640625,
|
||
|
|
"learning_rate": 4.700807357667952e-05,
|
||
|
|
"loss": 2.7622,
|
||
|
|
"step": 711
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.712,
|
||
|
|
"grad_norm": 0.61328125,
|
||
|
|
"learning_rate": 4.6712372392927e-05,
|
||
|
|
"loss": 2.7514,
|
||
|
|
"step": 712
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.713,
|
||
|
|
"grad_norm": 0.6640625,
|
||
|
|
"learning_rate": 4.6417320502100316e-05,
|
||
|
|
"loss": 2.8732,
|
||
|
|
"step": 713
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.714,
|
||
|
|
"grad_norm": 0.6015625,
|
||
|
|
"learning_rate": 4.612292149931369e-05,
|
||
|
|
"loss": 2.8161,
|
||
|
|
"step": 714
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.715,
|
||
|
|
"grad_norm": 0.578125,
|
||
|
|
"learning_rate": 4.582917897172603e-05,
|
||
|
|
"loss": 2.7898,
|
||
|
|
"step": 715
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.716,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 4.5536096498497295e-05,
|
||
|
|
"loss": 2.7848,
|
||
|
|
"step": 716
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.717,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 4.524367765074499e-05,
|
||
|
|
"loss": 2.703,
|
||
|
|
"step": 717
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.718,
|
||
|
|
"grad_norm": 0.59375,
|
||
|
|
"learning_rate": 4.495192599150044e-05,
|
||
|
|
"loss": 2.7208,
|
||
|
|
"step": 718
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.719,
|
||
|
|
"grad_norm": 1.3671875,
|
||
|
|
"learning_rate": 4.46608450756656e-05,
|
||
|
|
"loss": 2.767,
|
||
|
|
"step": 719
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.72,
|
||
|
|
"grad_norm": 0.57421875,
|
||
|
|
"learning_rate": 4.437043844996952e-05,
|
||
|
|
"loss": 2.7353,
|
||
|
|
"step": 720
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.721,
|
||
|
|
"grad_norm": 0.640625,
|
||
|
|
"learning_rate": 4.4080709652925336e-05,
|
||
|
|
"loss": 2.7186,
|
||
|
|
"step": 721
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.722,
|
||
|
|
"grad_norm": 18.625,
|
||
|
|
"learning_rate": 4.379166221478697e-05,
|
||
|
|
"loss": 2.6783,
|
||
|
|
"step": 722
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.723,
|
||
|
|
"grad_norm": 0.6015625,
|
||
|
|
"learning_rate": 4.350329965750621e-05,
|
||
|
|
"loss": 2.718,
|
||
|
|
"step": 723
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.724,
|
||
|
|
"grad_norm": 0.64453125,
|
||
|
|
"learning_rate": 4.32156254946899e-05,
|
||
|
|
"loss": 2.6782,
|
||
|
|
"step": 724
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.725,
|
||
|
|
"grad_norm": 0.6484375,
|
||
|
|
"learning_rate": 4.2928643231556844e-05,
|
||
|
|
"loss": 2.7787,
|
||
|
|
"step": 725
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.726,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 4.264235636489542e-05,
|
||
|
|
"loss": 2.7355,
|
||
|
|
"step": 726
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.727,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 4.235676838302068e-05,
|
||
|
|
"loss": 2.8365,
|
||
|
|
"step": 727
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.728,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 4.207188276573214e-05,
|
||
|
|
"loss": 2.7514,
|
||
|
|
"step": 728
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.729,
|
||
|
|
"grad_norm": 0.60546875,
|
||
|
|
"learning_rate": 4.1787702984271074e-05,
|
||
|
|
"loss": 2.7733,
|
||
|
|
"step": 729
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.73,
|
||
|
|
"grad_norm": 1.8984375,
|
||
|
|
"learning_rate": 4.150423250127845e-05,
|
||
|
|
"loss": 2.809,
|
||
|
|
"step": 730
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.731,
|
||
|
|
"grad_norm": 0.5859375,
|
||
|
|
"learning_rate": 4.12214747707527e-05,
|
||
|
|
"loss": 2.7394,
|
||
|
|
"step": 731
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.732,
|
||
|
|
"grad_norm": 0.578125,
|
||
|
|
"learning_rate": 4.093943323800745e-05,
|
||
|
|
"loss": 2.6793,
|
||
|
|
"step": 732
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.733,
|
||
|
|
"grad_norm": 0.5625,
|
||
|
|
"learning_rate": 4.065811133962987e-05,
|
||
|
|
"loss": 2.7069,
|
||
|
|
"step": 733
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.734,
|
||
|
|
"grad_norm": 0.57421875,
|
||
|
|
"learning_rate": 4.037751250343841e-05,
|
||
|
|
"loss": 2.6702,
|
||
|
|
"step": 734
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.735,
|
||
|
|
"grad_norm": 1.109375,
|
||
|
|
"learning_rate": 4.009764014844143e-05,
|
||
|
|
"loss": 2.8871,
|
||
|
|
"step": 735
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.736,
|
||
|
|
"grad_norm": 0.62890625,
|
||
|
|
"learning_rate": 3.981849768479517e-05,
|
||
|
|
"loss": 2.7682,
|
||
|
|
"step": 736
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.737,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 3.954008851376252e-05,
|
||
|
|
"loss": 2.7024,
|
||
|
|
"step": 737
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.738,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 3.9262416027671356e-05,
|
||
|
|
"loss": 2.6956,
|
||
|
|
"step": 738
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.739,
|
||
|
|
"grad_norm": 0.6640625,
|
||
|
|
"learning_rate": 3.8985483609873244e-05,
|
||
|
|
"loss": 2.8372,
|
||
|
|
"step": 739
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.74,
|
||
|
|
"grad_norm": 0.5703125,
|
||
|
|
"learning_rate": 3.8709294634702376e-05,
|
||
|
|
"loss": 2.7524,
|
||
|
|
"step": 740
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.741,
|
||
|
|
"grad_norm": 0.59375,
|
||
|
|
"learning_rate": 3.843385246743417e-05,
|
||
|
|
"loss": 2.8114,
|
||
|
|
"step": 741
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.742,
|
||
|
|
"grad_norm": 0.60546875,
|
||
|
|
"learning_rate": 3.8159160464244606e-05,
|
||
|
|
"loss": 2.8749,
|
||
|
|
"step": 742
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.743,
|
||
|
|
"grad_norm": 0.578125,
|
||
|
|
"learning_rate": 3.788522197216897e-05,
|
||
|
|
"loss": 2.7037,
|
||
|
|
"step": 743
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.744,
|
||
|
|
"grad_norm": 0.55078125,
|
||
|
|
"learning_rate": 3.7612040329061405e-05,
|
||
|
|
"loss": 2.7504,
|
||
|
|
"step": 744
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.745,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 3.733961886355398e-05,
|
||
|
|
"loss": 2.7219,
|
||
|
|
"step": 745
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.746,
|
||
|
|
"grad_norm": 0.60546875,
|
||
|
|
"learning_rate": 3.7067960895016275e-05,
|
||
|
|
"loss": 2.741,
|
||
|
|
"step": 746
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.747,
|
||
|
|
"grad_norm": 0.5625,
|
||
|
|
"learning_rate": 3.679706973351491e-05,
|
||
|
|
"loss": 2.7085,
|
||
|
|
"step": 747
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.748,
|
||
|
|
"grad_norm": 0.59375,
|
||
|
|
"learning_rate": 3.6526948679773257e-05,
|
||
|
|
"loss": 2.7705,
|
||
|
|
"step": 748
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.749,
|
||
|
|
"grad_norm": 0.58984375,
|
||
|
|
"learning_rate": 3.6257601025131026e-05,
|
||
|
|
"loss": 2.7403,
|
||
|
|
"step": 749
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.75,
|
||
|
|
"grad_norm": 0.9375,
|
||
|
|
"learning_rate": 3.5989030051504434e-05,
|
||
|
|
"loss": 2.7612,
|
||
|
|
"step": 750
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.75,
|
||
|
|
"eval_loss": 2.7113237380981445,
|
||
|
|
"eval_runtime": 31.7342,
|
||
|
|
"eval_samples_per_second": 15.756,
|
||
|
|
"eval_steps_per_second": 2.647,
|
||
|
|
"step": 750
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.751,
|
||
|
|
"grad_norm": 0.59375,
|
||
|
|
"learning_rate": 3.5721239031346066e-05,
|
||
|
|
"loss": 2.8165,
|
||
|
|
"step": 751
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.752,
|
||
|
|
"grad_norm": 0.5625,
|
||
|
|
"learning_rate": 3.545423122760493e-05,
|
||
|
|
"loss": 2.6717,
|
||
|
|
"step": 752
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.753,
|
||
|
|
"grad_norm": 0.578125,
|
||
|
|
"learning_rate": 3.518800989368691e-05,
|
||
|
|
"loss": 2.5895,
|
||
|
|
"step": 753
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.754,
|
||
|
|
"grad_norm": 0.5625,
|
||
|
|
"learning_rate": 3.492257827341492e-05,
|
||
|
|
"loss": 2.7452,
|
||
|
|
"step": 754
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.755,
|
||
|
|
"grad_norm": 0.5625,
|
||
|
|
"learning_rate": 3.465793960098945e-05,
|
||
|
|
"loss": 2.7445,
|
||
|
|
"step": 755
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.756,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 3.439409710094929e-05,
|
||
|
|
"loss": 2.805,
|
||
|
|
"step": 756
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.757,
|
||
|
|
"grad_norm": 0.640625,
|
||
|
|
"learning_rate": 3.413105398813195e-05,
|
||
|
|
"loss": 2.8297,
|
||
|
|
"step": 757
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.758,
|
||
|
|
"grad_norm": 0.55078125,
|
||
|
|
"learning_rate": 3.386881346763483e-05,
|
||
|
|
"loss": 2.6497,
|
||
|
|
"step": 758
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.759,
|
||
|
|
"grad_norm": 0.56640625,
|
||
|
|
"learning_rate": 3.360737873477584e-05,
|
||
|
|
"loss": 2.7854,
|
||
|
|
"step": 759
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.76,
|
||
|
|
"grad_norm": 0.58203125,
|
||
|
|
"learning_rate": 3.334675297505476e-05,
|
||
|
|
"loss": 2.7295,
|
||
|
|
"step": 760
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.761,
|
||
|
|
"grad_norm": 0.578125,
|
||
|
|
"learning_rate": 3.308693936411421e-05,
|
||
|
|
"loss": 2.7563,
|
||
|
|
"step": 761
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.762,
|
||
|
|
"grad_norm": 0.5859375,
|
||
|
|
"learning_rate": 3.2827941067700996e-05,
|
||
|
|
"loss": 2.8956,
|
||
|
|
"step": 762
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.763,
|
||
|
|
"grad_norm": 0.76953125,
|
||
|
|
"learning_rate": 3.2569761241627696e-05,
|
||
|
|
"loss": 2.7614,
|
||
|
|
"step": 763
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.764,
|
||
|
|
"grad_norm": 0.578125,
|
||
|
|
"learning_rate": 3.231240303173394e-05,
|
||
|
|
"loss": 2.7303,
|
||
|
|
"step": 764
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.765,
|
||
|
|
"grad_norm": 0.58984375,
|
||
|
|
"learning_rate": 3.205586957384838e-05,
|
||
|
|
"loss": 2.7826,
|
||
|
|
"step": 765
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.766,
|
||
|
|
"grad_norm": 1.3359375,
|
||
|
|
"learning_rate": 3.1800163993750166e-05,
|
||
|
|
"loss": 2.7078,
|
||
|
|
"step": 766
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.767,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 3.154528940713113e-05,
|
||
|
|
"loss": 2.6754,
|
||
|
|
"step": 767
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.768,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 3.129124891955771e-05,
|
||
|
|
"loss": 2.7744,
|
||
|
|
"step": 768
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.769,
|
||
|
|
"grad_norm": 0.5625,
|
||
|
|
"learning_rate": 3.103804562643302e-05,
|
||
|
|
"loss": 2.653,
|
||
|
|
"step": 769
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.77,
|
||
|
|
"grad_norm": 0.546875,
|
||
|
|
"learning_rate": 3.078568261295933e-05,
|
||
|
|
"loss": 2.7278,
|
||
|
|
"step": 770
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.771,
|
||
|
|
"grad_norm": 0.57421875,
|
||
|
|
"learning_rate": 3.053416295410026e-05,
|
||
|
|
"loss": 2.7533,
|
||
|
|
"step": 771
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.772,
|
||
|
|
"grad_norm": 0.60546875,
|
||
|
|
"learning_rate": 3.0283489714543556e-05,
|
||
|
|
"loss": 2.7639,
|
||
|
|
"step": 772
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.773,
|
||
|
|
"grad_norm": 0.578125,
|
||
|
|
"learning_rate": 3.0033665948663448e-05,
|
||
|
|
"loss": 2.7951,
|
||
|
|
"step": 773
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.774,
|
||
|
|
"grad_norm": 0.6015625,
|
||
|
|
"learning_rate": 2.9784694700483762e-05,
|
||
|
|
"loss": 2.7375,
|
||
|
|
"step": 774
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.775,
|
||
|
|
"grad_norm": 0.55859375,
|
||
|
|
"learning_rate": 2.953657900364053e-05,
|
||
|
|
"loss": 2.7481,
|
||
|
|
"step": 775
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.776,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 2.9289321881345254e-05,
|
||
|
|
"loss": 2.7437,
|
||
|
|
"step": 776
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.777,
|
||
|
|
"grad_norm": 0.58984375,
|
||
|
|
"learning_rate": 2.904292634634793e-05,
|
||
|
|
"loss": 2.7637,
|
||
|
|
"step": 777
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.778,
|
||
|
|
"grad_norm": 0.546875,
|
||
|
|
"learning_rate": 2.879739540090036e-05,
|
||
|
|
"loss": 2.7252,
|
||
|
|
"step": 778
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.779,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 2.8552732036719687e-05,
|
||
|
|
"loss": 2.5999,
|
||
|
|
"step": 779
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.78,
|
||
|
|
"grad_norm": 0.5625,
|
||
|
|
"learning_rate": 2.8308939234951726e-05,
|
||
|
|
"loss": 2.8094,
|
||
|
|
"step": 780
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.781,
|
||
|
|
"grad_norm": 0.58984375,
|
||
|
|
"learning_rate": 2.8066019966134904e-05,
|
||
|
|
"loss": 2.7139,
|
||
|
|
"step": 781
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.782,
|
||
|
|
"grad_norm": 1.46875,
|
||
|
|
"learning_rate": 2.7823977190163786e-05,
|
||
|
|
"loss": 2.8623,
|
||
|
|
"step": 782
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.783,
|
||
|
|
"grad_norm": 0.52734375,
|
||
|
|
"learning_rate": 2.7582813856253275e-05,
|
||
|
|
"loss": 2.7122,
|
||
|
|
"step": 783
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.784,
|
||
|
|
"grad_norm": 0.55078125,
|
||
|
|
"learning_rate": 2.734253290290242e-05,
|
||
|
|
"loss": 2.7772,
|
||
|
|
"step": 784
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.785,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 2.7103137257858868e-05,
|
||
|
|
"loss": 2.7612,
|
||
|
|
"step": 785
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.786,
|
||
|
|
"grad_norm": 0.5625,
|
||
|
|
"learning_rate": 2.6864629838082956e-05,
|
||
|
|
"loss": 2.7303,
|
||
|
|
"step": 786
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.787,
|
||
|
|
"grad_norm": 0.546875,
|
||
|
|
"learning_rate": 2.6627013549712355e-05,
|
||
|
|
"loss": 2.6967,
|
||
|
|
"step": 787
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.788,
|
||
|
|
"grad_norm": 0.5234375,
|
||
|
|
"learning_rate": 2.639029128802657e-05,
|
||
|
|
"loss": 2.675,
|
||
|
|
"step": 788
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.789,
|
||
|
|
"grad_norm": 0.52734375,
|
||
|
|
"learning_rate": 2.615446593741161e-05,
|
||
|
|
"loss": 2.771,
|
||
|
|
"step": 789
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.79,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 2.5919540371325e-05,
|
||
|
|
"loss": 2.7169,
|
||
|
|
"step": 790
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.791,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 2.5685517452260567e-05,
|
||
|
|
"loss": 2.6604,
|
||
|
|
"step": 791
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.792,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 2.5452400031713785e-05,
|
||
|
|
"loss": 2.6749,
|
||
|
|
"step": 792
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.793,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 2.5220190950146827e-05,
|
||
|
|
"loss": 2.6759,
|
||
|
|
"step": 793
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.794,
|
||
|
|
"grad_norm": 0.57421875,
|
||
|
|
"learning_rate": 2.4988893036954043e-05,
|
||
|
|
"loss": 2.5863,
|
||
|
|
"step": 794
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.795,
|
||
|
|
"grad_norm": 0.55078125,
|
||
|
|
"learning_rate": 2.4758509110427575e-05,
|
||
|
|
"loss": 2.7512,
|
||
|
|
"step": 795
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.796,
|
||
|
|
"grad_norm": 0.59765625,
|
||
|
|
"learning_rate": 2.45290419777228e-05,
|
||
|
|
"loss": 2.7619,
|
||
|
|
"step": 796
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.797,
|
||
|
|
"grad_norm": 0.74609375,
|
||
|
|
"learning_rate": 2.4300494434824373e-05,
|
||
|
|
"loss": 2.6975,
|
||
|
|
"step": 797
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.798,
|
||
|
|
"grad_norm": 0.56640625,
|
||
|
|
"learning_rate": 2.407286926651192e-05,
|
||
|
|
"loss": 2.6837,
|
||
|
|
"step": 798
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.799,
|
||
|
|
"grad_norm": 0.5234375,
|
||
|
|
"learning_rate": 2.3846169246326343e-05,
|
||
|
|
"loss": 2.7131,
|
||
|
|
"step": 799
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.8,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 2.362039713653581e-05,
|
||
|
|
"loss": 2.7429,
|
||
|
|
"step": 800
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.801,
|
||
|
|
"grad_norm": 0.578125,
|
||
|
|
"learning_rate": 2.339555568810221e-05,
|
||
|
|
"loss": 2.6673,
|
||
|
|
"step": 801
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.802,
|
||
|
|
"grad_norm": 0.57421875,
|
||
|
|
"learning_rate": 2.3171647640647687e-05,
|
||
|
|
"loss": 2.6941,
|
||
|
|
"step": 802
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.803,
|
||
|
|
"grad_norm": 0.6484375,
|
||
|
|
"learning_rate": 2.2948675722421086e-05,
|
||
|
|
"loss": 2.735,
|
||
|
|
"step": 803
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.804,
|
||
|
|
"grad_norm": 0.79296875,
|
||
|
|
"learning_rate": 2.2726642650264895e-05,
|
||
|
|
"loss": 2.7529,
|
||
|
|
"step": 804
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.805,
|
||
|
|
"grad_norm": 0.55859375,
|
||
|
|
"learning_rate": 2.2505551129582047e-05,
|
||
|
|
"loss": 2.6125,
|
||
|
|
"step": 805
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.806,
|
||
|
|
"grad_norm": 0.56640625,
|
||
|
|
"learning_rate": 2.2285403854302912e-05,
|
||
|
|
"loss": 2.8059,
|
||
|
|
"step": 806
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.807,
|
||
|
|
"grad_norm": 0.546875,
|
||
|
|
"learning_rate": 2.2066203506852566e-05,
|
||
|
|
"loss": 2.6916,
|
||
|
|
"step": 807
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.808,
|
||
|
|
"grad_norm": 0.7265625,
|
||
|
|
"learning_rate": 2.1847952758118117e-05,
|
||
|
|
"loss": 2.7839,
|
||
|
|
"step": 808
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.809,
|
||
|
|
"grad_norm": 0.58984375,
|
||
|
|
"learning_rate": 2.163065426741603e-05,
|
||
|
|
"loss": 2.7451,
|
||
|
|
"step": 809
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.81,
|
||
|
|
"grad_norm": 0.53515625,
|
||
|
|
"learning_rate": 2.1414310682459802e-05,
|
||
|
|
"loss": 2.7282,
|
||
|
|
"step": 810
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.811,
|
||
|
|
"grad_norm": 0.57421875,
|
||
|
|
"learning_rate": 2.119892463932781e-05,
|
||
|
|
"loss": 2.6989,
|
||
|
|
"step": 811
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.812,
|
||
|
|
"grad_norm": 0.5859375,
|
||
|
|
"learning_rate": 2.098449876243096e-05,
|
||
|
|
"loss": 2.6681,
|
||
|
|
"step": 812
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.813,
|
||
|
|
"grad_norm": 0.70703125,
|
||
|
|
"learning_rate": 2.0771035664480942e-05,
|
||
|
|
"loss": 2.7579,
|
||
|
|
"step": 813
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.814,
|
||
|
|
"grad_norm": 0.57421875,
|
||
|
|
"learning_rate": 2.0558537946458177e-05,
|
||
|
|
"loss": 2.6096,
|
||
|
|
"step": 814
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.815,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 2.0347008197580374e-05,
|
||
|
|
"loss": 2.5568,
|
||
|
|
"step": 815
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.816,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 2.013644899527074e-05,
|
||
|
|
"loss": 2.7679,
|
||
|
|
"step": 816
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.817,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 1.9926862905126665e-05,
|
||
|
|
"loss": 2.751,
|
||
|
|
"step": 817
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.818,
|
||
|
|
"grad_norm": 0.51953125,
|
||
|
|
"learning_rate": 1.9718252480888566e-05,
|
||
|
|
"loss": 2.7757,
|
||
|
|
"step": 818
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.819,
|
||
|
|
"grad_norm": 0.58203125,
|
||
|
|
"learning_rate": 1.9510620264408596e-05,
|
||
|
|
"loss": 2.6745,
|
||
|
|
"step": 819
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.82,
|
||
|
|
"grad_norm": 0.58203125,
|
||
|
|
"learning_rate": 1.930396878561983e-05,
|
||
|
|
"loss": 2.6728,
|
||
|
|
"step": 820
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.821,
|
||
|
|
"grad_norm": 0.5859375,
|
||
|
|
"learning_rate": 1.9098300562505266e-05,
|
||
|
|
"loss": 2.6893,
|
||
|
|
"step": 821
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.822,
|
||
|
|
"grad_norm": 0.58203125,
|
||
|
|
"learning_rate": 1.8893618101067355e-05,
|
||
|
|
"loss": 2.8697,
|
||
|
|
"step": 822
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.823,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 1.8689923895297245e-05,
|
||
|
|
"loss": 2.7268,
|
||
|
|
"step": 823
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.824,
|
||
|
|
"grad_norm": 0.578125,
|
||
|
|
"learning_rate": 1.848722042714457e-05,
|
||
|
|
"loss": 2.7974,
|
||
|
|
"step": 824
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.825,
|
||
|
|
"grad_norm": 0.53515625,
|
||
|
|
"learning_rate": 1.8285510166487152e-05,
|
||
|
|
"loss": 2.691,
|
||
|
|
"step": 825
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.826,
|
||
|
|
"grad_norm": 0.5859375,
|
||
|
|
"learning_rate": 1.808479557110081e-05,
|
||
|
|
"loss": 2.6877,
|
||
|
|
"step": 826
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.827,
|
||
|
|
"grad_norm": 0.53125,
|
||
|
|
"learning_rate": 1.78850790866296e-05,
|
||
|
|
"loss": 2.6971,
|
||
|
|
"step": 827
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.828,
|
||
|
|
"grad_norm": 0.5625,
|
||
|
|
"learning_rate": 1.7686363146555805e-05,
|
||
|
|
"loss": 2.8624,
|
||
|
|
"step": 828
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.829,
|
||
|
|
"grad_norm": 2.421875,
|
||
|
|
"learning_rate": 1.7488650172170496e-05,
|
||
|
|
"loss": 2.6911,
|
||
|
|
"step": 829
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.83,
|
||
|
|
"grad_norm": 0.5859375,
|
||
|
|
"learning_rate": 1.7291942572543807e-05,
|
||
|
|
"loss": 2.6905,
|
||
|
|
"step": 830
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.831,
|
||
|
|
"grad_norm": 0.58984375,
|
||
|
|
"learning_rate": 1.7096242744495837e-05,
|
||
|
|
"loss": 2.7459,
|
||
|
|
"step": 831
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.832,
|
||
|
|
"grad_norm": 0.6875,
|
||
|
|
"learning_rate": 1.690155307256719e-05,
|
||
|
|
"loss": 2.738,
|
||
|
|
"step": 832
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.833,
|
||
|
|
"grad_norm": 0.578125,
|
||
|
|
"learning_rate": 1.6707875928990058e-05,
|
||
|
|
"loss": 2.7459,
|
||
|
|
"step": 833
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.834,
|
||
|
|
"grad_norm": 0.5859375,
|
||
|
|
"learning_rate": 1.6515213673659357e-05,
|
||
|
|
"loss": 2.7432,
|
||
|
|
"step": 834
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.835,
|
||
|
|
"grad_norm": 0.55078125,
|
||
|
|
"learning_rate": 1.632356865410384e-05,
|
||
|
|
"loss": 2.6985,
|
||
|
|
"step": 835
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.836,
|
||
|
|
"grad_norm": 0.58203125,
|
||
|
|
"learning_rate": 1.6132943205457606e-05,
|
||
|
|
"loss": 2.8408,
|
||
|
|
"step": 836
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.837,
|
||
|
|
"grad_norm": 0.53515625,
|
||
|
|
"learning_rate": 1.5943339650431576e-05,
|
||
|
|
"loss": 2.662,
|
||
|
|
"step": 837
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.838,
|
||
|
|
"grad_norm": 0.53515625,
|
||
|
|
"learning_rate": 1.5754760299285252e-05,
|
||
|
|
"loss": 2.85,
|
||
|
|
"step": 838
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.839,
|
||
|
|
"grad_norm": 0.5390625,
|
||
|
|
"learning_rate": 1.5567207449798515e-05,
|
||
|
|
"loss": 2.6535,
|
||
|
|
"step": 839
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.84,
|
||
|
|
"grad_norm": 0.55078125,
|
||
|
|
"learning_rate": 1.538068338724361e-05,
|
||
|
|
"loss": 2.7144,
|
||
|
|
"step": 840
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.841,
|
||
|
|
"grad_norm": 0.53515625,
|
||
|
|
"learning_rate": 1.5195190384357404e-05,
|
||
|
|
"loss": 2.8185,
|
||
|
|
"step": 841
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.842,
|
||
|
|
"grad_norm": 1.0703125,
|
||
|
|
"learning_rate": 1.5010730701313625e-05,
|
||
|
|
"loss": 2.7978,
|
||
|
|
"step": 842
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.843,
|
||
|
|
"grad_norm": 0.546875,
|
||
|
|
"learning_rate": 1.4827306585695234e-05,
|
||
|
|
"loss": 2.7138,
|
||
|
|
"step": 843
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.844,
|
||
|
|
"grad_norm": 0.890625,
|
||
|
|
"learning_rate": 1.4644920272467244e-05,
|
||
|
|
"loss": 2.7772,
|
||
|
|
"step": 844
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.845,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 1.4463573983949341e-05,
|
||
|
|
"loss": 2.7598,
|
||
|
|
"step": 845
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.846,
|
||
|
|
"grad_norm": 0.6953125,
|
||
|
|
"learning_rate": 1.4283269929788779e-05,
|
||
|
|
"loss": 2.7201,
|
||
|
|
"step": 846
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.847,
|
||
|
|
"grad_norm": 0.51953125,
|
||
|
|
"learning_rate": 1.4104010306933557e-05,
|
||
|
|
"loss": 2.8091,
|
||
|
|
"step": 847
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.848,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 1.3925797299605647e-05,
|
||
|
|
"loss": 2.7008,
|
||
|
|
"step": 848
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.849,
|
||
|
|
"grad_norm": 0.52734375,
|
||
|
|
"learning_rate": 1.3748633079274253e-05,
|
||
|
|
"loss": 2.655,
|
||
|
|
"step": 849
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.85,
|
||
|
|
"grad_norm": 0.57421875,
|
||
|
|
"learning_rate": 1.3572519804629536e-05,
|
||
|
|
"loss": 2.6479,
|
||
|
|
"step": 850
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.851,
|
||
|
|
"grad_norm": 0.5625,
|
||
|
|
"learning_rate": 1.339745962155613e-05,
|
||
|
|
"loss": 2.8183,
|
||
|
|
"step": 851
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.852,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 1.3223454663107172e-05,
|
||
|
|
"loss": 2.7465,
|
||
|
|
"step": 852
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.853,
|
||
|
|
"grad_norm": 0.6015625,
|
||
|
|
"learning_rate": 1.30505070494781e-05,
|
||
|
|
"loss": 2.7232,
|
||
|
|
"step": 853
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.854,
|
||
|
|
"grad_norm": 0.5078125,
|
||
|
|
"learning_rate": 1.2878618887981064e-05,
|
||
|
|
"loss": 2.7987,
|
||
|
|
"step": 854
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.855,
|
||
|
|
"grad_norm": 0.52734375,
|
||
|
|
"learning_rate": 1.2707792273019048e-05,
|
||
|
|
"loss": 2.7005,
|
||
|
|
"step": 855
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.856,
|
||
|
|
"grad_norm": 0.55078125,
|
||
|
|
"learning_rate": 1.2538029286060426e-05,
|
||
|
|
"loss": 2.7156,
|
||
|
|
"step": 856
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.857,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 1.2369331995613665e-05,
|
||
|
|
"loss": 2.757,
|
||
|
|
"step": 857
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.858,
|
||
|
|
"grad_norm": 0.91015625,
|
||
|
|
"learning_rate": 1.2201702457201947e-05,
|
||
|
|
"loss": 2.6486,
|
||
|
|
"step": 858
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.859,
|
||
|
|
"grad_norm": 0.5703125,
|
||
|
|
"learning_rate": 1.2035142713338366e-05,
|
||
|
|
"loss": 2.7237,
|
||
|
|
"step": 859
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.86,
|
||
|
|
"grad_norm": 1.125,
|
||
|
|
"learning_rate": 1.1869654793500784e-05,
|
||
|
|
"loss": 2.6579,
|
||
|
|
"step": 860
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.861,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 1.1705240714107302e-05,
|
||
|
|
"loss": 2.719,
|
||
|
|
"step": 861
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.862,
|
||
|
|
"grad_norm": 0.51171875,
|
||
|
|
"learning_rate": 1.1541902478491606e-05,
|
||
|
|
"loss": 2.6551,
|
||
|
|
"step": 862
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.863,
|
||
|
|
"grad_norm": 0.52734375,
|
||
|
|
"learning_rate": 1.1379642076878527e-05,
|
||
|
|
"loss": 2.6851,
|
||
|
|
"step": 863
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.864,
|
||
|
|
"grad_norm": 0.5390625,
|
||
|
|
"learning_rate": 1.1218461486359877e-05,
|
||
|
|
"loss": 2.7538,
|
||
|
|
"step": 864
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.865,
|
||
|
|
"grad_norm": 0.58984375,
|
||
|
|
"learning_rate": 1.1058362670870249e-05,
|
||
|
|
"loss": 2.693,
|
||
|
|
"step": 865
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.866,
|
||
|
|
"grad_norm": 0.578125,
|
||
|
|
"learning_rate": 1.0899347581163221e-05,
|
||
|
|
"loss": 2.7243,
|
||
|
|
"step": 866
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.867,
|
||
|
|
"grad_norm": 0.546875,
|
||
|
|
"learning_rate": 1.0741418154787442e-05,
|
||
|
|
"loss": 2.6959,
|
||
|
|
"step": 867
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.868,
|
||
|
|
"grad_norm": 0.5703125,
|
||
|
|
"learning_rate": 1.0584576316063188e-05,
|
||
|
|
"loss": 2.7305,
|
||
|
|
"step": 868
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.869,
|
||
|
|
"grad_norm": 0.57421875,
|
||
|
|
"learning_rate": 1.042882397605871e-05,
|
||
|
|
"loss": 2.7807,
|
||
|
|
"step": 869
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.87,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 1.0274163032567163e-05,
|
||
|
|
"loss": 2.7001,
|
||
|
|
"step": 870
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.871,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 1.0120595370083318e-05,
|
||
|
|
"loss": 2.7309,
|
||
|
|
"step": 871
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.872,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 9.968122859780648e-06,
|
||
|
|
"loss": 2.6748,
|
||
|
|
"step": 872
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.873,
|
||
|
|
"grad_norm": 0.58203125,
|
||
|
|
"learning_rate": 9.816747359488632e-06,
|
||
|
|
"loss": 2.7391,
|
||
|
|
"step": 873
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.874,
|
||
|
|
"grad_norm": 0.5625,
|
||
|
|
"learning_rate": 9.666470713669918e-06,
|
||
|
|
"loss": 2.8011,
|
||
|
|
"step": 874
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.875,
|
||
|
|
"grad_norm": 1.21875,
|
||
|
|
"learning_rate": 9.517294753398064e-06,
|
||
|
|
"loss": 2.6658,
|
||
|
|
"step": 875
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.876,
|
||
|
|
"grad_norm": 0.546875,
|
||
|
|
"learning_rate": 9.369221296335006e-06,
|
||
|
|
"loss": 2.7542,
|
||
|
|
"step": 876
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.877,
|
||
|
|
"grad_norm": 0.5390625,
|
||
|
|
"learning_rate": 9.222252146709142e-06,
|
||
|
|
"loss": 2.6932,
|
||
|
|
"step": 877
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.878,
|
||
|
|
"grad_norm": 0.51953125,
|
||
|
|
"learning_rate": 9.076389095293148e-06,
|
||
|
|
"loss": 2.7678,
|
||
|
|
"step": 878
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.879,
|
||
|
|
"grad_norm": 0.53515625,
|
||
|
|
"learning_rate": 8.931633919382298e-06,
|
||
|
|
"loss": 2.6873,
|
||
|
|
"step": 879
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.88,
|
||
|
|
"grad_norm": 0.52734375,
|
||
|
|
"learning_rate": 8.787988382772705e-06,
|
||
|
|
"loss": 2.812,
|
||
|
|
"step": 880
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.881,
|
||
|
|
"grad_norm": 0.57421875,
|
||
|
|
"learning_rate": 8.645454235739903e-06,
|
||
|
|
"loss": 2.6715,
|
||
|
|
"step": 881
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.882,
|
||
|
|
"grad_norm": 0.53125,
|
||
|
|
"learning_rate": 8.504033215017527e-06,
|
||
|
|
"loss": 2.7504,
|
||
|
|
"step": 882
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.883,
|
||
|
|
"grad_norm": 0.515625,
|
||
|
|
"learning_rate": 8.363727043776038e-06,
|
||
|
|
"loss": 2.7513,
|
||
|
|
"step": 883
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.884,
|
||
|
|
"grad_norm": 0.51171875,
|
||
|
|
"learning_rate": 8.224537431601886e-06,
|
||
|
|
"loss": 2.8052,
|
||
|
|
"step": 884
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.885,
|
||
|
|
"grad_norm": 0.53125,
|
||
|
|
"learning_rate": 8.086466074476563e-06,
|
||
|
|
"loss": 2.7368,
|
||
|
|
"step": 885
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.886,
|
||
|
|
"grad_norm": 0.55859375,
|
||
|
|
"learning_rate": 7.949514654755962e-06,
|
||
|
|
"loss": 2.6757,
|
||
|
|
"step": 886
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.887,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 7.81368484114996e-06,
|
||
|
|
"loss": 2.703,
|
||
|
|
"step": 887
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.888,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 7.67897828870191e-06,
|
||
|
|
"loss": 2.7158,
|
||
|
|
"step": 888
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.889,
|
||
|
|
"grad_norm": 0.55078125,
|
||
|
|
"learning_rate": 7.545396638768698e-06,
|
||
|
|
"loss": 2.7778,
|
||
|
|
"step": 889
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.89,
|
||
|
|
"grad_norm": 0.5078125,
|
||
|
|
"learning_rate": 7.412941519000527e-06,
|
||
|
|
"loss": 2.6752,
|
||
|
|
"step": 890
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.891,
|
||
|
|
"grad_norm": 1.59375,
|
||
|
|
"learning_rate": 7.281614543321269e-06,
|
||
|
|
"loss": 2.7462,
|
||
|
|
"step": 891
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.892,
|
||
|
|
"grad_norm": 0.515625,
|
||
|
|
"learning_rate": 7.151417311908648e-06,
|
||
|
|
"loss": 2.6856,
|
||
|
|
"step": 892
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.893,
|
||
|
|
"grad_norm": 0.546875,
|
||
|
|
"learning_rate": 7.022351411174866e-06,
|
||
|
|
"loss": 2.7136,
|
||
|
|
"step": 893
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.894,
|
||
|
|
"grad_norm": 0.55859375,
|
||
|
|
"learning_rate": 6.894418413747183e-06,
|
||
|
|
"loss": 2.7652,
|
||
|
|
"step": 894
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.895,
|
||
|
|
"grad_norm": 0.5390625,
|
||
|
|
"learning_rate": 6.767619878448783e-06,
|
||
|
|
"loss": 2.687,
|
||
|
|
"step": 895
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.896,
|
||
|
|
"grad_norm": 0.609375,
|
||
|
|
"learning_rate": 6.6419573502798374e-06,
|
||
|
|
"loss": 2.8021,
|
||
|
|
"step": 896
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.897,
|
||
|
|
"grad_norm": 0.53125,
|
||
|
|
"learning_rate": 6.517432360398556e-06,
|
||
|
|
"loss": 2.7689,
|
||
|
|
"step": 897
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.898,
|
||
|
|
"grad_norm": 0.5859375,
|
||
|
|
"learning_rate": 6.394046426102674e-06,
|
||
|
|
"loss": 2.7352,
|
||
|
|
"step": 898
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.899,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 6.2718010508108545e-06,
|
||
|
|
"loss": 2.7,
|
||
|
|
"step": 899
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.9,
|
||
|
|
"grad_norm": 0.55859375,
|
||
|
|
"learning_rate": 6.1506977240444074e-06,
|
||
|
|
"loss": 2.6995,
|
||
|
|
"step": 900
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.901,
|
||
|
|
"grad_norm": 0.5390625,
|
||
|
|
"learning_rate": 6.030737921409169e-06,
|
||
|
|
"loss": 2.6935,
|
||
|
|
"step": 901
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.902,
|
||
|
|
"grad_norm": 0.5234375,
|
||
|
|
"learning_rate": 5.911923104577455e-06,
|
||
|
|
"loss": 2.711,
|
||
|
|
"step": 902
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.903,
|
||
|
|
"grad_norm": 0.5078125,
|
||
|
|
"learning_rate": 5.7942547212703315e-06,
|
||
|
|
"loss": 2.7778,
|
||
|
|
"step": 903
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.904,
|
||
|
|
"grad_norm": 0.53125,
|
||
|
|
"learning_rate": 5.6777342052399045e-06,
|
||
|
|
"loss": 2.7373,
|
||
|
|
"step": 904
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.905,
|
||
|
|
"grad_norm": 0.515625,
|
||
|
|
"learning_rate": 5.562362976251901e-06,
|
||
|
|
"loss": 2.7644,
|
||
|
|
"step": 905
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.906,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 5.448142440068316e-06,
|
||
|
|
"loss": 2.6753,
|
||
|
|
"step": 906
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.907,
|
||
|
|
"grad_norm": 0.8515625,
|
||
|
|
"learning_rate": 5.335073988430372e-06,
|
||
|
|
"loss": 2.7881,
|
||
|
|
"step": 907
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.908,
|
||
|
|
"grad_norm": 0.5234375,
|
||
|
|
"learning_rate": 5.223158999041444e-06,
|
||
|
|
"loss": 2.8192,
|
||
|
|
"step": 908
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.909,
|
||
|
|
"grad_norm": 0.578125,
|
||
|
|
"learning_rate": 5.1123988355503475e-06,
|
||
|
|
"loss": 2.7979,
|
||
|
|
"step": 909
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.91,
|
||
|
|
"grad_norm": 0.578125,
|
||
|
|
"learning_rate": 5.002794847534764e-06,
|
||
|
|
"loss": 2.7629,
|
||
|
|
"step": 910
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.911,
|
||
|
|
"grad_norm": 0.53515625,
|
||
|
|
"learning_rate": 4.8943483704846475e-06,
|
||
|
|
"loss": 2.7704,
|
||
|
|
"step": 911
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.912,
|
||
|
|
"grad_norm": 0.55859375,
|
||
|
|
"learning_rate": 4.7870607257861415e-06,
|
||
|
|
"loss": 2.6811,
|
||
|
|
"step": 912
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.913,
|
||
|
|
"grad_norm": 0.515625,
|
||
|
|
"learning_rate": 4.680933220705308e-06,
|
||
|
|
"loss": 2.7518,
|
||
|
|
"step": 913
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.914,
|
||
|
|
"grad_norm": 0.53125,
|
||
|
|
"learning_rate": 4.575967148372317e-06,
|
||
|
|
"loss": 2.7595,
|
||
|
|
"step": 914
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.915,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 4.4721637877656375e-06,
|
||
|
|
"loss": 2.7555,
|
||
|
|
"step": 915
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.916,
|
||
|
|
"grad_norm": 0.546875,
|
||
|
|
"learning_rate": 4.369524403696457e-06,
|
||
|
|
"loss": 2.728,
|
||
|
|
"step": 916
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.917,
|
||
|
|
"grad_norm": 0.546875,
|
||
|
|
"learning_rate": 4.268050246793276e-06,
|
||
|
|
"loss": 2.7264,
|
||
|
|
"step": 917
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.918,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 4.167742553486675e-06,
|
||
|
|
"loss": 2.7391,
|
||
|
|
"step": 918
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.919,
|
||
|
|
"grad_norm": 0.53515625,
|
||
|
|
"learning_rate": 4.068602545994249e-06,
|
||
|
|
"loss": 2.7269,
|
||
|
|
"step": 919
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.92,
|
||
|
|
"grad_norm": 0.51953125,
|
||
|
|
"learning_rate": 3.970631432305694e-06,
|
||
|
|
"loss": 2.7273,
|
||
|
|
"step": 920
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.921,
|
||
|
|
"grad_norm": 0.53515625,
|
||
|
|
"learning_rate": 3.873830406168111e-06,
|
||
|
|
"loss": 2.8126,
|
||
|
|
"step": 921
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.922,
|
||
|
|
"grad_norm": 2.171875,
|
||
|
|
"learning_rate": 3.7782006470714616e-06,
|
||
|
|
"loss": 2.7532,
|
||
|
|
"step": 922
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.923,
|
||
|
|
"grad_norm": 0.53125,
|
||
|
|
"learning_rate": 3.68374332023419e-06,
|
||
|
|
"loss": 2.7923,
|
||
|
|
"step": 923
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.924,
|
||
|
|
"grad_norm": 0.52734375,
|
||
|
|
"learning_rate": 3.5904595765890005e-06,
|
||
|
|
"loss": 2.7534,
|
||
|
|
"step": 924
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.925,
|
||
|
|
"grad_norm": 0.5390625,
|
||
|
|
"learning_rate": 3.4983505527688586e-06,
|
||
|
|
"loss": 2.7706,
|
||
|
|
"step": 925
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.926,
|
||
|
|
"grad_norm": 0.5390625,
|
||
|
|
"learning_rate": 3.40741737109318e-06,
|
||
|
|
"loss": 2.6821,
|
||
|
|
"step": 926
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.927,
|
||
|
|
"grad_norm": 0.56640625,
|
||
|
|
"learning_rate": 3.3176611395540626e-06,
|
||
|
|
"loss": 2.6992,
|
||
|
|
"step": 927
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.928,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 3.2290829518028862e-06,
|
||
|
|
"loss": 2.7818,
|
||
|
|
"step": 928
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.929,
|
||
|
|
"grad_norm": 0.56640625,
|
||
|
|
"learning_rate": 3.1416838871368924e-06,
|
||
|
|
"loss": 2.7538,
|
||
|
|
"step": 929
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.93,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 3.0554650104861136e-06,
|
||
|
|
"loss": 2.7118,
|
||
|
|
"step": 930
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.931,
|
||
|
|
"grad_norm": 0.55859375,
|
||
|
|
"learning_rate": 2.970427372400353e-06,
|
||
|
|
"loss": 2.7252,
|
||
|
|
"step": 931
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.932,
|
||
|
|
"grad_norm": 0.546875,
|
||
|
|
"learning_rate": 2.8865720090364034e-06,
|
||
|
|
"loss": 2.8308,
|
||
|
|
"step": 932
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.933,
|
||
|
|
"grad_norm": 0.53125,
|
||
|
|
"learning_rate": 2.8038999421453826e-06,
|
||
|
|
"loss": 2.6437,
|
||
|
|
"step": 933
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.934,
|
||
|
|
"grad_norm": 0.50390625,
|
||
|
|
"learning_rate": 2.7224121790603517e-06,
|
||
|
|
"loss": 2.764,
|
||
|
|
"step": 934
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.935,
|
||
|
|
"grad_norm": 0.5625,
|
||
|
|
"learning_rate": 2.6421097126839712e-06,
|
||
|
|
"loss": 2.6876,
|
||
|
|
"step": 935
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.936,
|
||
|
|
"grad_norm": 0.55078125,
|
||
|
|
"learning_rate": 2.5629935214764865e-06,
|
||
|
|
"loss": 2.8505,
|
||
|
|
"step": 936
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.937,
|
||
|
|
"grad_norm": 0.53515625,
|
||
|
|
"learning_rate": 2.4850645694436736e-06,
|
||
|
|
"loss": 2.7782,
|
||
|
|
"step": 937
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.938,
|
||
|
|
"grad_norm": 1.0625,
|
||
|
|
"learning_rate": 2.4083238061252567e-06,
|
||
|
|
"loss": 2.6892,
|
||
|
|
"step": 938
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.939,
|
||
|
|
"grad_norm": 0.52734375,
|
||
|
|
"learning_rate": 2.332772166583208e-06,
|
||
|
|
"loss": 2.6548,
|
||
|
|
"step": 939
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.94,
|
||
|
|
"grad_norm": 0.5625,
|
||
|
|
"learning_rate": 2.2584105713904125e-06,
|
||
|
|
"loss": 2.7622,
|
||
|
|
"step": 940
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.941,
|
||
|
|
"grad_norm": 0.5625,
|
||
|
|
"learning_rate": 2.1852399266194314e-06,
|
||
|
|
"loss": 2.7183,
|
||
|
|
"step": 941
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.942,
|
||
|
|
"grad_norm": 0.5234375,
|
||
|
|
"learning_rate": 2.1132611238315003e-06,
|
||
|
|
"loss": 2.7956,
|
||
|
|
"step": 942
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.943,
|
||
|
|
"grad_norm": 0.50390625,
|
||
|
|
"learning_rate": 2.0424750400655947e-06,
|
||
|
|
"loss": 2.6357,
|
||
|
|
"step": 943
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.944,
|
||
|
|
"grad_norm": 0.55859375,
|
||
|
|
"learning_rate": 1.9728825378278246e-06,
|
||
|
|
"loss": 2.6503,
|
||
|
|
"step": 944
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.945,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 1.904484465080847e-06,
|
||
|
|
"loss": 2.6973,
|
||
|
|
"step": 945
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.946,
|
||
|
|
"grad_norm": 0.546875,
|
||
|
|
"learning_rate": 1.8372816552336026e-06,
|
||
|
|
"loss": 2.7222,
|
||
|
|
"step": 946
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.947,
|
||
|
|
"grad_norm": 0.515625,
|
||
|
|
"learning_rate": 1.771274927131139e-06,
|
||
|
|
"loss": 2.6586,
|
||
|
|
"step": 947
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.948,
|
||
|
|
"grad_norm": 0.53125,
|
||
|
|
"learning_rate": 1.706465085044584e-06,
|
||
|
|
"loss": 2.657,
|
||
|
|
"step": 948
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.949,
|
||
|
|
"grad_norm": 0.55078125,
|
||
|
|
"learning_rate": 1.6428529186614195e-06,
|
||
|
|
"loss": 2.7474,
|
||
|
|
"step": 949
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.95,
|
||
|
|
"grad_norm": 0.578125,
|
||
|
|
"learning_rate": 1.580439203075812e-06,
|
||
|
|
"loss": 2.6946,
|
||
|
|
"step": 950
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.951,
|
||
|
|
"grad_norm": 0.5234375,
|
||
|
|
"learning_rate": 1.5192246987791981e-06,
|
||
|
|
"loss": 2.6957,
|
||
|
|
"step": 951
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.952,
|
||
|
|
"grad_norm": 0.53125,
|
||
|
|
"learning_rate": 1.4592101516509914e-06,
|
||
|
|
"loss": 2.7191,
|
||
|
|
"step": 952
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.953,
|
||
|
|
"grad_norm": 0.5234375,
|
||
|
|
"learning_rate": 1.400396292949513e-06,
|
||
|
|
"loss": 2.6774,
|
||
|
|
"step": 953
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.954,
|
||
|
|
"grad_norm": 0.640625,
|
||
|
|
"learning_rate": 1.3427838393030633e-06,
|
||
|
|
"loss": 2.785,
|
||
|
|
"step": 954
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.955,
|
||
|
|
"grad_norm": 0.578125,
|
||
|
|
"learning_rate": 1.2863734927012095e-06,
|
||
|
|
"loss": 2.7072,
|
||
|
|
"step": 955
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.956,
|
||
|
|
"grad_norm": 0.55078125,
|
||
|
|
"learning_rate": 1.231165940486234e-06,
|
||
|
|
"loss": 2.7511,
|
||
|
|
"step": 956
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.957,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 1.1771618553447216e-06,
|
||
|
|
"loss": 2.7826,
|
||
|
|
"step": 957
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.958,
|
||
|
|
"grad_norm": 5.4375,
|
||
|
|
"learning_rate": 1.1243618952994195e-06,
|
||
|
|
"loss": 2.7722,
|
||
|
|
"step": 958
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.959,
|
||
|
|
"grad_norm": 0.50390625,
|
||
|
|
"learning_rate": 1.0727667037011668e-06,
|
||
|
|
"loss": 2.6715,
|
||
|
|
"step": 959
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.96,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 1.0223769092211012e-06,
|
||
|
|
"loss": 2.7709,
|
||
|
|
"step": 960
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.961,
|
||
|
|
"grad_norm": 0.5390625,
|
||
|
|
"learning_rate": 9.731931258429638e-07,
|
||
|
|
"loss": 2.7166,
|
||
|
|
"step": 961
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.962,
|
||
|
|
"grad_norm": 0.53515625,
|
||
|
|
"learning_rate": 9.252159528556403e-07,
|
||
|
|
"loss": 2.7073,
|
||
|
|
"step": 962
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.963,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 8.784459748458318e-07,
|
||
|
|
"loss": 2.6861,
|
||
|
|
"step": 963
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.964,
|
||
|
|
"grad_norm": 0.515625,
|
||
|
|
"learning_rate": 8.328837616909613e-07,
|
||
|
|
"loss": 2.7969,
|
||
|
|
"step": 964
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.965,
|
||
|
|
"grad_norm": 0.53125,
|
||
|
|
"learning_rate": 7.885298685522235e-07,
|
||
|
|
"loss": 2.6999,
|
||
|
|
"step": 965
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.966,
|
||
|
|
"grad_norm": 0.56640625,
|
||
|
|
"learning_rate": 7.453848358678017e-07,
|
||
|
|
"loss": 2.7762,
|
||
|
|
"step": 966
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.967,
|
||
|
|
"grad_norm": 0.55078125,
|
||
|
|
"learning_rate": 7.034491893463058e-07,
|
||
|
|
"loss": 2.7231,
|
||
|
|
"step": 967
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.968,
|
||
|
|
"grad_norm": 0.515625,
|
||
|
|
"learning_rate": 6.627234399603555e-07,
|
||
|
|
"loss": 2.6857,
|
||
|
|
"step": 968
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.969,
|
||
|
|
"grad_norm": 2.28125,
|
||
|
|
"learning_rate": 6.232080839403631e-07,
|
||
|
|
"loss": 2.707,
|
||
|
|
"step": 969
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.97,
|
||
|
|
"grad_norm": 0.56640625,
|
||
|
|
"learning_rate": 5.849036027684606e-07,
|
||
|
|
"loss": 2.7359,
|
||
|
|
"step": 970
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.971,
|
||
|
|
"grad_norm": 0.55078125,
|
||
|
|
"learning_rate": 5.478104631726711e-07,
|
||
|
|
"loss": 2.6725,
|
||
|
|
"step": 971
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.972,
|
||
|
|
"grad_norm": 0.53125,
|
||
|
|
"learning_rate": 5.119291171211793e-07,
|
||
|
|
"loss": 2.762,
|
||
|
|
"step": 972
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.973,
|
||
|
|
"grad_norm": 0.546875,
|
||
|
|
"learning_rate": 4.772600018168816e-07,
|
||
|
|
"loss": 2.7416,
|
||
|
|
"step": 973
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.974,
|
||
|
|
"grad_norm": 0.53125,
|
||
|
|
"learning_rate": 4.438035396920004e-07,
|
||
|
|
"loss": 2.7811,
|
||
|
|
"step": 974
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.975,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 4.115601384029666e-07,
|
||
|
|
"loss": 2.7719,
|
||
|
|
"step": 975
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.976,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 3.805301908254455e-07,
|
||
|
|
"loss": 2.7054,
|
||
|
|
"step": 976
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.977,
|
||
|
|
"grad_norm": 0.546875,
|
||
|
|
"learning_rate": 3.50714075049563e-07,
|
||
|
|
"loss": 2.8006,
|
||
|
|
"step": 977
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.978,
|
||
|
|
"grad_norm": 0.58203125,
|
||
|
|
"learning_rate": 3.2211215437528694e-07,
|
||
|
|
"loss": 2.7788,
|
||
|
|
"step": 978
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.979,
|
||
|
|
"grad_norm": 0.53125,
|
||
|
|
"learning_rate": 2.947247773079753e-07,
|
||
|
|
"loss": 2.6917,
|
||
|
|
"step": 979
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.98,
|
||
|
|
"grad_norm": 0.5625,
|
||
|
|
"learning_rate": 2.685522775541904e-07,
|
||
|
|
"loss": 2.7921,
|
||
|
|
"step": 980
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.981,
|
||
|
|
"grad_norm": 0.58984375,
|
||
|
|
"learning_rate": 2.4359497401758024e-07,
|
||
|
|
"loss": 2.6928,
|
||
|
|
"step": 981
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.982,
|
||
|
|
"grad_norm": 0.515625,
|
||
|
|
"learning_rate": 2.1985317079500356e-07,
|
||
|
|
"loss": 2.8196,
|
||
|
|
"step": 982
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.983,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 1.973271571728441e-07,
|
||
|
|
"loss": 2.6882,
|
||
|
|
"step": 983
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.984,
|
||
|
|
"grad_norm": 0.546875,
|
||
|
|
"learning_rate": 1.7601720762346897e-07,
|
||
|
|
"loss": 2.7932,
|
||
|
|
"step": 984
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.985,
|
||
|
|
"grad_norm": 0.79296875,
|
||
|
|
"learning_rate": 1.5592358180189782e-07,
|
||
|
|
"loss": 2.7674,
|
||
|
|
"step": 985
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.986,
|
||
|
|
"grad_norm": 0.5,
|
||
|
|
"learning_rate": 1.3704652454261668e-07,
|
||
|
|
"loss": 2.7227,
|
||
|
|
"step": 986
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.987,
|
||
|
|
"grad_norm": 0.5859375,
|
||
|
|
"learning_rate": 1.193862658566025e-07,
|
||
|
|
"loss": 2.6739,
|
||
|
|
"step": 987
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.988,
|
||
|
|
"grad_norm": 0.52734375,
|
||
|
|
"learning_rate": 1.0294302092853647e-07,
|
||
|
|
"loss": 2.731,
|
||
|
|
"step": 988
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.989,
|
||
|
|
"grad_norm": 0.578125,
|
||
|
|
"learning_rate": 8.771699011416168e-08,
|
||
|
|
"loss": 2.8035,
|
||
|
|
"step": 989
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.99,
|
||
|
|
"grad_norm": 0.53515625,
|
||
|
|
"learning_rate": 7.370835893788508e-08,
|
||
|
|
"loss": 2.7294,
|
||
|
|
"step": 990
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.991,
|
||
|
|
"grad_norm": 0.55078125,
|
||
|
|
"learning_rate": 6.09172980904238e-08,
|
||
|
|
"loss": 2.8871,
|
||
|
|
"step": 991
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.992,
|
||
|
|
"grad_norm": 0.5390625,
|
||
|
|
"learning_rate": 4.934396342684e-08,
|
||
|
|
"loss": 2.7629,
|
||
|
|
"step": 992
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.993,
|
||
|
|
"grad_norm": 0.5859375,
|
||
|
|
"learning_rate": 3.898849596456478e-08,
|
||
|
|
"loss": 2.7404,
|
||
|
|
"step": 993
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.994,
|
||
|
|
"grad_norm": 0.5546875,
|
||
|
|
"learning_rate": 2.985102188168831e-08,
|
||
|
|
"loss": 2.7835,
|
||
|
|
"step": 994
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.995,
|
||
|
|
"grad_norm": 0.54296875,
|
||
|
|
"learning_rate": 2.193165251545004e-08,
|
||
|
|
"loss": 2.7149,
|
||
|
|
"step": 995
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.996,
|
||
|
|
"grad_norm": 0.515625,
|
||
|
|
"learning_rate": 1.5230484360873044e-08,
|
||
|
|
"loss": 2.7078,
|
||
|
|
"step": 996
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.997,
|
||
|
|
"grad_norm": 0.53515625,
|
||
|
|
"learning_rate": 9.747599069576119e-09,
|
||
|
|
"loss": 2.6312,
|
||
|
|
"step": 997
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.998,
|
||
|
|
"grad_norm": 0.5078125,
|
||
|
|
"learning_rate": 5.483063448785686e-09,
|
||
|
|
"loss": 2.7117,
|
||
|
|
"step": 998
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 0.999,
|
||
|
|
"grad_norm": 0.53515625,
|
||
|
|
"learning_rate": 2.4369294605253166e-09,
|
||
|
|
"loss": 2.7865,
|
||
|
|
"step": 999
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 1.0,
|
||
|
|
"grad_norm": 0.5859375,
|
||
|
|
"learning_rate": 6.092342209607083e-10,
|
||
|
|
"loss": 2.7628,
|
||
|
|
"step": 1000
|
||
|
|
},
|
||
|
|
{
|
||
|
|
"epoch": 1.0,
|
||
|
|
"eval_loss": 2.704561948776245,
|
||
|
|
"eval_runtime": 31.803,
|
||
|
|
"eval_samples_per_second": 15.722,
|
||
|
|
"eval_steps_per_second": 2.641,
|
||
|
|
"step": 1000
|
||
|
|
}
|
||
|
|
],
|
||
|
|
"logging_steps": 1,
|
||
|
|
"max_steps": 1000,
|
||
|
|
"num_input_tokens_seen": 0,
|
||
|
|
"num_train_epochs": 9223372036854775807,
|
||
|
|
"save_steps": 1000,
|
||
|
|
"stateful_callbacks": {
|
||
|
|
"TrainerControl": {
|
||
|
|
"args": {
|
||
|
|
"should_epoch_stop": false,
|
||
|
|
"should_evaluate": false,
|
||
|
|
"should_log": false,
|
||
|
|
"should_save": true,
|
||
|
|
"should_training_stop": true
|
||
|
|
},
|
||
|
|
"attributes": {}
|
||
|
|
}
|
||
|
|
},
|
||
|
|
"total_flos": 1.818479221014528e+18,
|
||
|
|
"train_batch_size": 6,
|
||
|
|
"trial_name": null,
|
||
|
|
"trial_params": null
|
||
|
|
}
|