7404 lines
144 KiB
JSON
7404 lines
144 KiB
JSON
{
|
|
"best_metric": null,
|
|
"best_model_checkpoint": null,
|
|
"epoch": 3.9933665008291874,
|
|
"eval_steps": 61,
|
|
"global_step": 1204,
|
|
"is_hyper_param_search": false,
|
|
"is_local_process_zero": true,
|
|
"is_world_process_zero": true,
|
|
"log_history": [
|
|
{
|
|
"epoch": 0.0,
|
|
"learning_rate": 0.0,
|
|
"loss": 1.5877,
|
|
"step": 1
|
|
},
|
|
{
|
|
"epoch": 0.0,
|
|
"eval_loss": 1.244094729423523,
|
|
"eval_runtime": 17.5479,
|
|
"eval_samples_per_second": 130.044,
|
|
"eval_steps_per_second": 5.471,
|
|
"step": 1
|
|
},
|
|
{
|
|
"epoch": 0.01,
|
|
"learning_rate": 6.000000000000001e-08,
|
|
"loss": 1.2159,
|
|
"step": 2
|
|
},
|
|
{
|
|
"epoch": 0.01,
|
|
"learning_rate": 1.2000000000000002e-07,
|
|
"loss": 1.2301,
|
|
"step": 3
|
|
},
|
|
{
|
|
"epoch": 0.01,
|
|
"learning_rate": 1.8e-07,
|
|
"loss": 1.1937,
|
|
"step": 4
|
|
},
|
|
{
|
|
"epoch": 0.02,
|
|
"learning_rate": 2.4000000000000003e-07,
|
|
"loss": 1.184,
|
|
"step": 5
|
|
},
|
|
{
|
|
"epoch": 0.02,
|
|
"learning_rate": 3.0000000000000004e-07,
|
|
"loss": 1.1365,
|
|
"step": 6
|
|
},
|
|
{
|
|
"epoch": 0.02,
|
|
"learning_rate": 3.6e-07,
|
|
"loss": 1.0685,
|
|
"step": 7
|
|
},
|
|
{
|
|
"epoch": 0.03,
|
|
"learning_rate": 4.2000000000000006e-07,
|
|
"loss": 1.1337,
|
|
"step": 8
|
|
},
|
|
{
|
|
"epoch": 0.03,
|
|
"learning_rate": 4.800000000000001e-07,
|
|
"loss": 1.0577,
|
|
"step": 9
|
|
},
|
|
{
|
|
"epoch": 0.03,
|
|
"learning_rate": 5.4e-07,
|
|
"loss": 1.0703,
|
|
"step": 10
|
|
},
|
|
{
|
|
"epoch": 0.04,
|
|
"learning_rate": 6.000000000000001e-07,
|
|
"loss": 1.0909,
|
|
"step": 11
|
|
},
|
|
{
|
|
"epoch": 0.04,
|
|
"learning_rate": 6.6e-07,
|
|
"loss": 1.0383,
|
|
"step": 12
|
|
},
|
|
{
|
|
"epoch": 0.04,
|
|
"learning_rate": 7.2e-07,
|
|
"loss": 1.0356,
|
|
"step": 13
|
|
},
|
|
{
|
|
"epoch": 0.05,
|
|
"learning_rate": 7.8e-07,
|
|
"loss": 1.0192,
|
|
"step": 14
|
|
},
|
|
{
|
|
"epoch": 0.05,
|
|
"learning_rate": 8.400000000000001e-07,
|
|
"loss": 1.0148,
|
|
"step": 15
|
|
},
|
|
{
|
|
"epoch": 0.05,
|
|
"learning_rate": 9e-07,
|
|
"loss": 1.0145,
|
|
"step": 16
|
|
},
|
|
{
|
|
"epoch": 0.06,
|
|
"learning_rate": 9.600000000000001e-07,
|
|
"loss": 0.9817,
|
|
"step": 17
|
|
},
|
|
{
|
|
"epoch": 0.06,
|
|
"learning_rate": 1.0200000000000002e-06,
|
|
"loss": 0.9746,
|
|
"step": 18
|
|
},
|
|
{
|
|
"epoch": 0.06,
|
|
"learning_rate": 1.08e-06,
|
|
"loss": 0.9888,
|
|
"step": 19
|
|
},
|
|
{
|
|
"epoch": 0.07,
|
|
"learning_rate": 1.14e-06,
|
|
"loss": 1.0165,
|
|
"step": 20
|
|
},
|
|
{
|
|
"epoch": 0.07,
|
|
"learning_rate": 1.2000000000000002e-06,
|
|
"loss": 0.9819,
|
|
"step": 21
|
|
},
|
|
{
|
|
"epoch": 0.07,
|
|
"learning_rate": 1.26e-06,
|
|
"loss": 0.9652,
|
|
"step": 22
|
|
},
|
|
{
|
|
"epoch": 0.08,
|
|
"learning_rate": 1.32e-06,
|
|
"loss": 0.9625,
|
|
"step": 23
|
|
},
|
|
{
|
|
"epoch": 0.08,
|
|
"learning_rate": 1.3800000000000001e-06,
|
|
"loss": 0.9734,
|
|
"step": 24
|
|
},
|
|
{
|
|
"epoch": 0.08,
|
|
"learning_rate": 1.44e-06,
|
|
"loss": 0.9506,
|
|
"step": 25
|
|
},
|
|
{
|
|
"epoch": 0.09,
|
|
"learning_rate": 1.5e-06,
|
|
"loss": 0.9716,
|
|
"step": 26
|
|
},
|
|
{
|
|
"epoch": 0.09,
|
|
"learning_rate": 1.56e-06,
|
|
"loss": 0.939,
|
|
"step": 27
|
|
},
|
|
{
|
|
"epoch": 0.09,
|
|
"learning_rate": 1.6200000000000002e-06,
|
|
"loss": 0.8999,
|
|
"step": 28
|
|
},
|
|
{
|
|
"epoch": 0.1,
|
|
"learning_rate": 1.6800000000000002e-06,
|
|
"loss": 0.9211,
|
|
"step": 29
|
|
},
|
|
{
|
|
"epoch": 0.1,
|
|
"learning_rate": 1.7399999999999999e-06,
|
|
"loss": 0.9174,
|
|
"step": 30
|
|
},
|
|
{
|
|
"epoch": 0.1,
|
|
"learning_rate": 1.8e-06,
|
|
"loss": 0.9036,
|
|
"step": 31
|
|
},
|
|
{
|
|
"epoch": 0.11,
|
|
"learning_rate": 1.86e-06,
|
|
"loss": 0.9197,
|
|
"step": 32
|
|
},
|
|
{
|
|
"epoch": 0.11,
|
|
"learning_rate": 1.9200000000000003e-06,
|
|
"loss": 0.9075,
|
|
"step": 33
|
|
},
|
|
{
|
|
"epoch": 0.11,
|
|
"learning_rate": 1.98e-06,
|
|
"loss": 0.9186,
|
|
"step": 34
|
|
},
|
|
{
|
|
"epoch": 0.12,
|
|
"learning_rate": 2.0400000000000004e-06,
|
|
"loss": 0.9153,
|
|
"step": 35
|
|
},
|
|
{
|
|
"epoch": 0.12,
|
|
"learning_rate": 2.1e-06,
|
|
"loss": 0.9065,
|
|
"step": 36
|
|
},
|
|
{
|
|
"epoch": 0.12,
|
|
"learning_rate": 2.16e-06,
|
|
"loss": 0.8857,
|
|
"step": 37
|
|
},
|
|
{
|
|
"epoch": 0.13,
|
|
"learning_rate": 2.22e-06,
|
|
"loss": 0.888,
|
|
"step": 38
|
|
},
|
|
{
|
|
"epoch": 0.13,
|
|
"learning_rate": 2.28e-06,
|
|
"loss": 0.8808,
|
|
"step": 39
|
|
},
|
|
{
|
|
"epoch": 0.13,
|
|
"learning_rate": 2.34e-06,
|
|
"loss": 0.8693,
|
|
"step": 40
|
|
},
|
|
{
|
|
"epoch": 0.14,
|
|
"learning_rate": 2.4000000000000003e-06,
|
|
"loss": 0.8923,
|
|
"step": 41
|
|
},
|
|
{
|
|
"epoch": 0.14,
|
|
"learning_rate": 2.4599999999999997e-06,
|
|
"loss": 0.8929,
|
|
"step": 42
|
|
},
|
|
{
|
|
"epoch": 0.14,
|
|
"learning_rate": 2.52e-06,
|
|
"loss": 0.8972,
|
|
"step": 43
|
|
},
|
|
{
|
|
"epoch": 0.15,
|
|
"learning_rate": 2.58e-06,
|
|
"loss": 0.8623,
|
|
"step": 44
|
|
},
|
|
{
|
|
"epoch": 0.15,
|
|
"learning_rate": 2.64e-06,
|
|
"loss": 0.8593,
|
|
"step": 45
|
|
},
|
|
{
|
|
"epoch": 0.15,
|
|
"learning_rate": 2.7e-06,
|
|
"loss": 0.8636,
|
|
"step": 46
|
|
},
|
|
{
|
|
"epoch": 0.16,
|
|
"learning_rate": 2.7600000000000003e-06,
|
|
"loss": 0.8686,
|
|
"step": 47
|
|
},
|
|
{
|
|
"epoch": 0.16,
|
|
"learning_rate": 2.82e-06,
|
|
"loss": 0.8819,
|
|
"step": 48
|
|
},
|
|
{
|
|
"epoch": 0.16,
|
|
"learning_rate": 2.88e-06,
|
|
"loss": 0.8765,
|
|
"step": 49
|
|
},
|
|
{
|
|
"epoch": 0.17,
|
|
"learning_rate": 2.9400000000000002e-06,
|
|
"loss": 0.8475,
|
|
"step": 50
|
|
},
|
|
{
|
|
"epoch": 0.17,
|
|
"learning_rate": 3e-06,
|
|
"loss": 0.8913,
|
|
"step": 51
|
|
},
|
|
{
|
|
"epoch": 0.17,
|
|
"learning_rate": 3.06e-06,
|
|
"loss": 0.8542,
|
|
"step": 52
|
|
},
|
|
{
|
|
"epoch": 0.18,
|
|
"learning_rate": 3.12e-06,
|
|
"loss": 0.8621,
|
|
"step": 53
|
|
},
|
|
{
|
|
"epoch": 0.18,
|
|
"learning_rate": 3.18e-06,
|
|
"loss": 0.8482,
|
|
"step": 54
|
|
},
|
|
{
|
|
"epoch": 0.18,
|
|
"learning_rate": 3.2400000000000003e-06,
|
|
"loss": 0.8436,
|
|
"step": 55
|
|
},
|
|
{
|
|
"epoch": 0.19,
|
|
"learning_rate": 3.3e-06,
|
|
"loss": 0.8483,
|
|
"step": 56
|
|
},
|
|
{
|
|
"epoch": 0.19,
|
|
"learning_rate": 3.3600000000000004e-06,
|
|
"loss": 0.8578,
|
|
"step": 57
|
|
},
|
|
{
|
|
"epoch": 0.19,
|
|
"learning_rate": 3.42e-06,
|
|
"loss": 0.8448,
|
|
"step": 58
|
|
},
|
|
{
|
|
"epoch": 0.2,
|
|
"learning_rate": 3.4799999999999997e-06,
|
|
"loss": 0.8409,
|
|
"step": 59
|
|
},
|
|
{
|
|
"epoch": 0.2,
|
|
"learning_rate": 3.54e-06,
|
|
"loss": 0.8577,
|
|
"step": 60
|
|
},
|
|
{
|
|
"epoch": 0.2,
|
|
"learning_rate": 3.6e-06,
|
|
"loss": 0.8481,
|
|
"step": 61
|
|
},
|
|
{
|
|
"epoch": 0.2,
|
|
"eval_loss": 0.847072184085846,
|
|
"eval_runtime": 17.615,
|
|
"eval_samples_per_second": 129.549,
|
|
"eval_steps_per_second": 5.45,
|
|
"step": 61
|
|
},
|
|
{
|
|
"epoch": 0.21,
|
|
"learning_rate": 3.66e-06,
|
|
"loss": 0.8365,
|
|
"step": 62
|
|
},
|
|
{
|
|
"epoch": 0.21,
|
|
"learning_rate": 3.72e-06,
|
|
"loss": 0.8379,
|
|
"step": 63
|
|
},
|
|
{
|
|
"epoch": 0.21,
|
|
"learning_rate": 3.7800000000000002e-06,
|
|
"loss": 0.8497,
|
|
"step": 64
|
|
},
|
|
{
|
|
"epoch": 0.22,
|
|
"learning_rate": 3.8400000000000005e-06,
|
|
"loss": 0.8284,
|
|
"step": 65
|
|
},
|
|
{
|
|
"epoch": 0.22,
|
|
"learning_rate": 3.9e-06,
|
|
"loss": 0.86,
|
|
"step": 66
|
|
},
|
|
{
|
|
"epoch": 0.22,
|
|
"learning_rate": 3.96e-06,
|
|
"loss": 0.8257,
|
|
"step": 67
|
|
},
|
|
{
|
|
"epoch": 0.23,
|
|
"learning_rate": 4.0200000000000005e-06,
|
|
"loss": 0.8296,
|
|
"step": 68
|
|
},
|
|
{
|
|
"epoch": 0.23,
|
|
"learning_rate": 4.080000000000001e-06,
|
|
"loss": 0.8163,
|
|
"step": 69
|
|
},
|
|
{
|
|
"epoch": 0.23,
|
|
"learning_rate": 4.14e-06,
|
|
"loss": 0.8516,
|
|
"step": 70
|
|
},
|
|
{
|
|
"epoch": 0.24,
|
|
"learning_rate": 4.2e-06,
|
|
"loss": 0.8297,
|
|
"step": 71
|
|
},
|
|
{
|
|
"epoch": 0.24,
|
|
"learning_rate": 4.26e-06,
|
|
"loss": 0.8288,
|
|
"step": 72
|
|
},
|
|
{
|
|
"epoch": 0.24,
|
|
"learning_rate": 4.32e-06,
|
|
"loss": 0.845,
|
|
"step": 73
|
|
},
|
|
{
|
|
"epoch": 0.25,
|
|
"learning_rate": 4.38e-06,
|
|
"loss": 0.829,
|
|
"step": 74
|
|
},
|
|
{
|
|
"epoch": 0.25,
|
|
"learning_rate": 4.44e-06,
|
|
"loss": 0.8348,
|
|
"step": 75
|
|
},
|
|
{
|
|
"epoch": 0.25,
|
|
"learning_rate": 4.5e-06,
|
|
"loss": 0.7995,
|
|
"step": 76
|
|
},
|
|
{
|
|
"epoch": 0.26,
|
|
"learning_rate": 4.56e-06,
|
|
"loss": 0.8406,
|
|
"step": 77
|
|
},
|
|
{
|
|
"epoch": 0.26,
|
|
"learning_rate": 4.62e-06,
|
|
"loss": 0.7984,
|
|
"step": 78
|
|
},
|
|
{
|
|
"epoch": 0.26,
|
|
"learning_rate": 4.68e-06,
|
|
"loss": 0.8122,
|
|
"step": 79
|
|
},
|
|
{
|
|
"epoch": 0.27,
|
|
"learning_rate": 4.74e-06,
|
|
"loss": 0.8225,
|
|
"step": 80
|
|
},
|
|
{
|
|
"epoch": 0.27,
|
|
"learning_rate": 4.800000000000001e-06,
|
|
"loss": 0.8467,
|
|
"step": 81
|
|
},
|
|
{
|
|
"epoch": 0.27,
|
|
"learning_rate": 4.86e-06,
|
|
"loss": 0.7995,
|
|
"step": 82
|
|
},
|
|
{
|
|
"epoch": 0.28,
|
|
"learning_rate": 4.9199999999999995e-06,
|
|
"loss": 0.8356,
|
|
"step": 83
|
|
},
|
|
{
|
|
"epoch": 0.28,
|
|
"learning_rate": 4.98e-06,
|
|
"loss": 0.8208,
|
|
"step": 84
|
|
},
|
|
{
|
|
"epoch": 0.28,
|
|
"learning_rate": 5.04e-06,
|
|
"loss": 0.8105,
|
|
"step": 85
|
|
},
|
|
{
|
|
"epoch": 0.29,
|
|
"learning_rate": 5.1e-06,
|
|
"loss": 0.8146,
|
|
"step": 86
|
|
},
|
|
{
|
|
"epoch": 0.29,
|
|
"learning_rate": 5.16e-06,
|
|
"loss": 0.7911,
|
|
"step": 87
|
|
},
|
|
{
|
|
"epoch": 0.29,
|
|
"learning_rate": 5.22e-06,
|
|
"loss": 0.8042,
|
|
"step": 88
|
|
},
|
|
{
|
|
"epoch": 0.3,
|
|
"learning_rate": 5.28e-06,
|
|
"loss": 0.8248,
|
|
"step": 89
|
|
},
|
|
{
|
|
"epoch": 0.3,
|
|
"learning_rate": 5.3400000000000005e-06,
|
|
"loss": 0.8032,
|
|
"step": 90
|
|
},
|
|
{
|
|
"epoch": 0.3,
|
|
"learning_rate": 5.4e-06,
|
|
"loss": 0.7966,
|
|
"step": 91
|
|
},
|
|
{
|
|
"epoch": 0.31,
|
|
"learning_rate": 5.46e-06,
|
|
"loss": 0.8026,
|
|
"step": 92
|
|
},
|
|
{
|
|
"epoch": 0.31,
|
|
"learning_rate": 5.5200000000000005e-06,
|
|
"loss": 0.8044,
|
|
"step": 93
|
|
},
|
|
{
|
|
"epoch": 0.31,
|
|
"learning_rate": 5.580000000000001e-06,
|
|
"loss": 0.8161,
|
|
"step": 94
|
|
},
|
|
{
|
|
"epoch": 0.32,
|
|
"learning_rate": 5.64e-06,
|
|
"loss": 0.8268,
|
|
"step": 95
|
|
},
|
|
{
|
|
"epoch": 0.32,
|
|
"learning_rate": 5.7e-06,
|
|
"loss": 0.8354,
|
|
"step": 96
|
|
},
|
|
{
|
|
"epoch": 0.32,
|
|
"learning_rate": 5.76e-06,
|
|
"loss": 0.8109,
|
|
"step": 97
|
|
},
|
|
{
|
|
"epoch": 0.33,
|
|
"learning_rate": 5.82e-06,
|
|
"loss": 0.7893,
|
|
"step": 98
|
|
},
|
|
{
|
|
"epoch": 0.33,
|
|
"learning_rate": 5.8800000000000005e-06,
|
|
"loss": 0.8184,
|
|
"step": 99
|
|
},
|
|
{
|
|
"epoch": 0.33,
|
|
"learning_rate": 5.94e-06,
|
|
"loss": 0.8148,
|
|
"step": 100
|
|
},
|
|
{
|
|
"epoch": 0.33,
|
|
"learning_rate": 6e-06,
|
|
"loss": 0.8211,
|
|
"step": 101
|
|
},
|
|
{
|
|
"epoch": 0.34,
|
|
"learning_rate": 5.994565217391305e-06,
|
|
"loss": 0.7684,
|
|
"step": 102
|
|
},
|
|
{
|
|
"epoch": 0.34,
|
|
"learning_rate": 5.9891304347826085e-06,
|
|
"loss": 0.8235,
|
|
"step": 103
|
|
},
|
|
{
|
|
"epoch": 0.34,
|
|
"learning_rate": 5.9836956521739135e-06,
|
|
"loss": 0.8006,
|
|
"step": 104
|
|
},
|
|
{
|
|
"epoch": 0.35,
|
|
"learning_rate": 5.978260869565218e-06,
|
|
"loss": 0.8169,
|
|
"step": 105
|
|
},
|
|
{
|
|
"epoch": 0.35,
|
|
"learning_rate": 5.972826086956522e-06,
|
|
"loss": 0.8,
|
|
"step": 106
|
|
},
|
|
{
|
|
"epoch": 0.35,
|
|
"learning_rate": 5.967391304347826e-06,
|
|
"loss": 0.8101,
|
|
"step": 107
|
|
},
|
|
{
|
|
"epoch": 0.36,
|
|
"learning_rate": 5.961956521739131e-06,
|
|
"loss": 0.7895,
|
|
"step": 108
|
|
},
|
|
{
|
|
"epoch": 0.36,
|
|
"learning_rate": 5.9565217391304344e-06,
|
|
"loss": 0.819,
|
|
"step": 109
|
|
},
|
|
{
|
|
"epoch": 0.36,
|
|
"learning_rate": 5.9510869565217395e-06,
|
|
"loss": 0.7907,
|
|
"step": 110
|
|
},
|
|
{
|
|
"epoch": 0.37,
|
|
"learning_rate": 5.945652173913044e-06,
|
|
"loss": 0.8062,
|
|
"step": 111
|
|
},
|
|
{
|
|
"epoch": 0.37,
|
|
"learning_rate": 5.940217391304348e-06,
|
|
"loss": 0.786,
|
|
"step": 112
|
|
},
|
|
{
|
|
"epoch": 0.37,
|
|
"learning_rate": 5.934782608695652e-06,
|
|
"loss": 0.8129,
|
|
"step": 113
|
|
},
|
|
{
|
|
"epoch": 0.38,
|
|
"learning_rate": 5.929347826086957e-06,
|
|
"loss": 0.7893,
|
|
"step": 114
|
|
},
|
|
{
|
|
"epoch": 0.38,
|
|
"learning_rate": 5.923913043478261e-06,
|
|
"loss": 0.7873,
|
|
"step": 115
|
|
},
|
|
{
|
|
"epoch": 0.38,
|
|
"learning_rate": 5.918478260869565e-06,
|
|
"loss": 0.7966,
|
|
"step": 116
|
|
},
|
|
{
|
|
"epoch": 0.39,
|
|
"learning_rate": 5.91304347826087e-06,
|
|
"loss": 0.797,
|
|
"step": 117
|
|
},
|
|
{
|
|
"epoch": 0.39,
|
|
"learning_rate": 5.907608695652174e-06,
|
|
"loss": 0.8147,
|
|
"step": 118
|
|
},
|
|
{
|
|
"epoch": 0.39,
|
|
"learning_rate": 5.902173913043479e-06,
|
|
"loss": 0.7911,
|
|
"step": 119
|
|
},
|
|
{
|
|
"epoch": 0.4,
|
|
"learning_rate": 5.896739130434783e-06,
|
|
"loss": 0.8058,
|
|
"step": 120
|
|
},
|
|
{
|
|
"epoch": 0.4,
|
|
"learning_rate": 5.891304347826087e-06,
|
|
"loss": 0.7975,
|
|
"step": 121
|
|
},
|
|
{
|
|
"epoch": 0.4,
|
|
"learning_rate": 5.885869565217391e-06,
|
|
"loss": 0.7841,
|
|
"step": 122
|
|
},
|
|
{
|
|
"epoch": 0.4,
|
|
"eval_loss": 0.8003882765769958,
|
|
"eval_runtime": 17.6189,
|
|
"eval_samples_per_second": 129.52,
|
|
"eval_steps_per_second": 5.449,
|
|
"step": 122
|
|
},
|
|
{
|
|
"epoch": 0.41,
|
|
"learning_rate": 5.880434782608696e-06,
|
|
"loss": 0.7799,
|
|
"step": 123
|
|
},
|
|
{
|
|
"epoch": 0.41,
|
|
"learning_rate": 5.875e-06,
|
|
"loss": 0.81,
|
|
"step": 124
|
|
},
|
|
{
|
|
"epoch": 0.41,
|
|
"learning_rate": 5.869565217391305e-06,
|
|
"loss": 0.7855,
|
|
"step": 125
|
|
},
|
|
{
|
|
"epoch": 0.42,
|
|
"learning_rate": 5.864130434782609e-06,
|
|
"loss": 0.7849,
|
|
"step": 126
|
|
},
|
|
{
|
|
"epoch": 0.42,
|
|
"learning_rate": 5.858695652173913e-06,
|
|
"loss": 0.8151,
|
|
"step": 127
|
|
},
|
|
{
|
|
"epoch": 0.42,
|
|
"learning_rate": 5.853260869565217e-06,
|
|
"loss": 0.7934,
|
|
"step": 128
|
|
},
|
|
{
|
|
"epoch": 0.43,
|
|
"learning_rate": 5.847826086956522e-06,
|
|
"loss": 0.7854,
|
|
"step": 129
|
|
},
|
|
{
|
|
"epoch": 0.43,
|
|
"learning_rate": 5.842391304347826e-06,
|
|
"loss": 0.7896,
|
|
"step": 130
|
|
},
|
|
{
|
|
"epoch": 0.43,
|
|
"learning_rate": 5.836956521739131e-06,
|
|
"loss": 0.7825,
|
|
"step": 131
|
|
},
|
|
{
|
|
"epoch": 0.44,
|
|
"learning_rate": 5.831521739130435e-06,
|
|
"loss": 0.8191,
|
|
"step": 132
|
|
},
|
|
{
|
|
"epoch": 0.44,
|
|
"learning_rate": 5.826086956521739e-06,
|
|
"loss": 0.8136,
|
|
"step": 133
|
|
},
|
|
{
|
|
"epoch": 0.44,
|
|
"learning_rate": 5.820652173913044e-06,
|
|
"loss": 0.8003,
|
|
"step": 134
|
|
},
|
|
{
|
|
"epoch": 0.45,
|
|
"learning_rate": 5.815217391304348e-06,
|
|
"loss": 0.7949,
|
|
"step": 135
|
|
},
|
|
{
|
|
"epoch": 0.45,
|
|
"learning_rate": 5.809782608695652e-06,
|
|
"loss": 0.7857,
|
|
"step": 136
|
|
},
|
|
{
|
|
"epoch": 0.45,
|
|
"learning_rate": 5.8043478260869565e-06,
|
|
"loss": 0.7895,
|
|
"step": 137
|
|
},
|
|
{
|
|
"epoch": 0.46,
|
|
"learning_rate": 5.798913043478261e-06,
|
|
"loss": 0.7935,
|
|
"step": 138
|
|
},
|
|
{
|
|
"epoch": 0.46,
|
|
"learning_rate": 5.793478260869565e-06,
|
|
"loss": 0.8038,
|
|
"step": 139
|
|
},
|
|
{
|
|
"epoch": 0.46,
|
|
"learning_rate": 5.78804347826087e-06,
|
|
"loss": 0.7798,
|
|
"step": 140
|
|
},
|
|
{
|
|
"epoch": 0.47,
|
|
"learning_rate": 5.782608695652174e-06,
|
|
"loss": 0.7824,
|
|
"step": 141
|
|
},
|
|
{
|
|
"epoch": 0.47,
|
|
"learning_rate": 5.777173913043478e-06,
|
|
"loss": 0.7733,
|
|
"step": 142
|
|
},
|
|
{
|
|
"epoch": 0.47,
|
|
"learning_rate": 5.771739130434783e-06,
|
|
"loss": 0.79,
|
|
"step": 143
|
|
},
|
|
{
|
|
"epoch": 0.48,
|
|
"learning_rate": 5.7663043478260875e-06,
|
|
"loss": 0.8024,
|
|
"step": 144
|
|
},
|
|
{
|
|
"epoch": 0.48,
|
|
"learning_rate": 5.760869565217392e-06,
|
|
"loss": 0.781,
|
|
"step": 145
|
|
},
|
|
{
|
|
"epoch": 0.48,
|
|
"learning_rate": 5.755434782608696e-06,
|
|
"loss": 0.7831,
|
|
"step": 146
|
|
},
|
|
{
|
|
"epoch": 0.49,
|
|
"learning_rate": 5.75e-06,
|
|
"loss": 0.7842,
|
|
"step": 147
|
|
},
|
|
{
|
|
"epoch": 0.49,
|
|
"learning_rate": 5.744565217391304e-06,
|
|
"loss": 0.7848,
|
|
"step": 148
|
|
},
|
|
{
|
|
"epoch": 0.49,
|
|
"learning_rate": 5.739130434782609e-06,
|
|
"loss": 0.7876,
|
|
"step": 149
|
|
},
|
|
{
|
|
"epoch": 0.5,
|
|
"learning_rate": 5.733695652173913e-06,
|
|
"loss": 0.7922,
|
|
"step": 150
|
|
},
|
|
{
|
|
"epoch": 0.5,
|
|
"learning_rate": 5.7282608695652176e-06,
|
|
"loss": 0.7728,
|
|
"step": 151
|
|
},
|
|
{
|
|
"epoch": 0.5,
|
|
"learning_rate": 5.722826086956522e-06,
|
|
"loss": 0.7838,
|
|
"step": 152
|
|
},
|
|
{
|
|
"epoch": 0.51,
|
|
"learning_rate": 5.717391304347826e-06,
|
|
"loss": 0.7843,
|
|
"step": 153
|
|
},
|
|
{
|
|
"epoch": 0.51,
|
|
"learning_rate": 5.71195652173913e-06,
|
|
"loss": 0.7721,
|
|
"step": 154
|
|
},
|
|
{
|
|
"epoch": 0.51,
|
|
"learning_rate": 5.706521739130435e-06,
|
|
"loss": 0.7946,
|
|
"step": 155
|
|
},
|
|
{
|
|
"epoch": 0.52,
|
|
"learning_rate": 5.701086956521739e-06,
|
|
"loss": 0.8002,
|
|
"step": 156
|
|
},
|
|
{
|
|
"epoch": 0.52,
|
|
"learning_rate": 5.6956521739130435e-06,
|
|
"loss": 0.7786,
|
|
"step": 157
|
|
},
|
|
{
|
|
"epoch": 0.52,
|
|
"learning_rate": 5.6902173913043485e-06,
|
|
"loss": 0.7848,
|
|
"step": 158
|
|
},
|
|
{
|
|
"epoch": 0.53,
|
|
"learning_rate": 5.684782608695652e-06,
|
|
"loss": 0.8104,
|
|
"step": 159
|
|
},
|
|
{
|
|
"epoch": 0.53,
|
|
"learning_rate": 5.679347826086957e-06,
|
|
"loss": 0.7767,
|
|
"step": 160
|
|
},
|
|
{
|
|
"epoch": 0.53,
|
|
"learning_rate": 5.673913043478261e-06,
|
|
"loss": 0.7822,
|
|
"step": 161
|
|
},
|
|
{
|
|
"epoch": 0.54,
|
|
"learning_rate": 5.668478260869565e-06,
|
|
"loss": 0.8001,
|
|
"step": 162
|
|
},
|
|
{
|
|
"epoch": 0.54,
|
|
"learning_rate": 5.663043478260869e-06,
|
|
"loss": 0.7854,
|
|
"step": 163
|
|
},
|
|
{
|
|
"epoch": 0.54,
|
|
"learning_rate": 5.6576086956521744e-06,
|
|
"loss": 0.7942,
|
|
"step": 164
|
|
},
|
|
{
|
|
"epoch": 0.55,
|
|
"learning_rate": 5.652173913043479e-06,
|
|
"loss": 0.8057,
|
|
"step": 165
|
|
},
|
|
{
|
|
"epoch": 0.55,
|
|
"learning_rate": 5.646739130434783e-06,
|
|
"loss": 0.7702,
|
|
"step": 166
|
|
},
|
|
{
|
|
"epoch": 0.55,
|
|
"learning_rate": 5.641304347826087e-06,
|
|
"loss": 0.7799,
|
|
"step": 167
|
|
},
|
|
{
|
|
"epoch": 0.56,
|
|
"learning_rate": 5.635869565217391e-06,
|
|
"loss": 0.7892,
|
|
"step": 168
|
|
},
|
|
{
|
|
"epoch": 0.56,
|
|
"learning_rate": 5.630434782608695e-06,
|
|
"loss": 0.7715,
|
|
"step": 169
|
|
},
|
|
{
|
|
"epoch": 0.56,
|
|
"learning_rate": 5.625e-06,
|
|
"loss": 0.7636,
|
|
"step": 170
|
|
},
|
|
{
|
|
"epoch": 0.57,
|
|
"learning_rate": 5.6195652173913045e-06,
|
|
"loss": 0.7875,
|
|
"step": 171
|
|
},
|
|
{
|
|
"epoch": 0.57,
|
|
"learning_rate": 5.614130434782609e-06,
|
|
"loss": 0.7824,
|
|
"step": 172
|
|
},
|
|
{
|
|
"epoch": 0.57,
|
|
"learning_rate": 5.608695652173914e-06,
|
|
"loss": 0.7769,
|
|
"step": 173
|
|
},
|
|
{
|
|
"epoch": 0.58,
|
|
"learning_rate": 5.603260869565217e-06,
|
|
"loss": 0.793,
|
|
"step": 174
|
|
},
|
|
{
|
|
"epoch": 0.58,
|
|
"learning_rate": 5.597826086956522e-06,
|
|
"loss": 0.7667,
|
|
"step": 175
|
|
},
|
|
{
|
|
"epoch": 0.58,
|
|
"learning_rate": 5.592391304347826e-06,
|
|
"loss": 0.7643,
|
|
"step": 176
|
|
},
|
|
{
|
|
"epoch": 0.59,
|
|
"learning_rate": 5.5869565217391305e-06,
|
|
"loss": 0.7718,
|
|
"step": 177
|
|
},
|
|
{
|
|
"epoch": 0.59,
|
|
"learning_rate": 5.581521739130435e-06,
|
|
"loss": 0.7583,
|
|
"step": 178
|
|
},
|
|
{
|
|
"epoch": 0.59,
|
|
"learning_rate": 5.57608695652174e-06,
|
|
"loss": 0.7722,
|
|
"step": 179
|
|
},
|
|
{
|
|
"epoch": 0.6,
|
|
"learning_rate": 5.570652173913043e-06,
|
|
"loss": 0.7739,
|
|
"step": 180
|
|
},
|
|
{
|
|
"epoch": 0.6,
|
|
"learning_rate": 5.565217391304348e-06,
|
|
"loss": 0.7785,
|
|
"step": 181
|
|
},
|
|
{
|
|
"epoch": 0.6,
|
|
"learning_rate": 5.559782608695652e-06,
|
|
"loss": 0.7653,
|
|
"step": 182
|
|
},
|
|
{
|
|
"epoch": 0.61,
|
|
"learning_rate": 5.554347826086956e-06,
|
|
"loss": 0.7927,
|
|
"step": 183
|
|
},
|
|
{
|
|
"epoch": 0.61,
|
|
"eval_loss": 0.7837424278259277,
|
|
"eval_runtime": 17.6126,
|
|
"eval_samples_per_second": 129.566,
|
|
"eval_steps_per_second": 5.451,
|
|
"step": 183
|
|
},
|
|
{
|
|
"epoch": 0.61,
|
|
"learning_rate": 5.548913043478261e-06,
|
|
"loss": 0.7651,
|
|
"step": 184
|
|
},
|
|
{
|
|
"epoch": 0.61,
|
|
"learning_rate": 5.543478260869566e-06,
|
|
"loss": 0.7851,
|
|
"step": 185
|
|
},
|
|
{
|
|
"epoch": 0.62,
|
|
"learning_rate": 5.53804347826087e-06,
|
|
"loss": 0.7797,
|
|
"step": 186
|
|
},
|
|
{
|
|
"epoch": 0.62,
|
|
"learning_rate": 5.532608695652174e-06,
|
|
"loss": 0.7893,
|
|
"step": 187
|
|
},
|
|
{
|
|
"epoch": 0.62,
|
|
"learning_rate": 5.527173913043479e-06,
|
|
"loss": 0.783,
|
|
"step": 188
|
|
},
|
|
{
|
|
"epoch": 0.63,
|
|
"learning_rate": 5.521739130434782e-06,
|
|
"loss": 0.7599,
|
|
"step": 189
|
|
},
|
|
{
|
|
"epoch": 0.63,
|
|
"learning_rate": 5.516304347826087e-06,
|
|
"loss": 0.7859,
|
|
"step": 190
|
|
},
|
|
{
|
|
"epoch": 0.63,
|
|
"learning_rate": 5.5108695652173915e-06,
|
|
"loss": 0.7765,
|
|
"step": 191
|
|
},
|
|
{
|
|
"epoch": 0.64,
|
|
"learning_rate": 5.505434782608696e-06,
|
|
"loss": 0.7577,
|
|
"step": 192
|
|
},
|
|
{
|
|
"epoch": 0.64,
|
|
"learning_rate": 5.5e-06,
|
|
"loss": 0.7743,
|
|
"step": 193
|
|
},
|
|
{
|
|
"epoch": 0.64,
|
|
"learning_rate": 5.494565217391305e-06,
|
|
"loss": 0.7792,
|
|
"step": 194
|
|
},
|
|
{
|
|
"epoch": 0.65,
|
|
"learning_rate": 5.489130434782608e-06,
|
|
"loss": 0.7652,
|
|
"step": 195
|
|
},
|
|
{
|
|
"epoch": 0.65,
|
|
"learning_rate": 5.483695652173913e-06,
|
|
"loss": 0.765,
|
|
"step": 196
|
|
},
|
|
{
|
|
"epoch": 0.65,
|
|
"learning_rate": 5.478260869565217e-06,
|
|
"loss": 0.7859,
|
|
"step": 197
|
|
},
|
|
{
|
|
"epoch": 0.66,
|
|
"learning_rate": 5.472826086956522e-06,
|
|
"loss": 0.7969,
|
|
"step": 198
|
|
},
|
|
{
|
|
"epoch": 0.66,
|
|
"learning_rate": 5.467391304347827e-06,
|
|
"loss": 0.7669,
|
|
"step": 199
|
|
},
|
|
{
|
|
"epoch": 0.66,
|
|
"learning_rate": 5.461956521739131e-06,
|
|
"loss": 0.7502,
|
|
"step": 200
|
|
},
|
|
{
|
|
"epoch": 0.67,
|
|
"learning_rate": 5.456521739130435e-06,
|
|
"loss": 0.7658,
|
|
"step": 201
|
|
},
|
|
{
|
|
"epoch": 0.67,
|
|
"learning_rate": 5.451086956521739e-06,
|
|
"loss": 0.7901,
|
|
"step": 202
|
|
},
|
|
{
|
|
"epoch": 0.67,
|
|
"learning_rate": 5.445652173913044e-06,
|
|
"loss": 0.7658,
|
|
"step": 203
|
|
},
|
|
{
|
|
"epoch": 0.68,
|
|
"learning_rate": 5.4402173913043475e-06,
|
|
"loss": 0.7589,
|
|
"step": 204
|
|
},
|
|
{
|
|
"epoch": 0.68,
|
|
"learning_rate": 5.4347826086956525e-06,
|
|
"loss": 0.7774,
|
|
"step": 205
|
|
},
|
|
{
|
|
"epoch": 0.68,
|
|
"learning_rate": 5.429347826086957e-06,
|
|
"loss": 0.7738,
|
|
"step": 206
|
|
},
|
|
{
|
|
"epoch": 0.69,
|
|
"learning_rate": 5.423913043478261e-06,
|
|
"loss": 0.801,
|
|
"step": 207
|
|
},
|
|
{
|
|
"epoch": 0.69,
|
|
"learning_rate": 5.418478260869565e-06,
|
|
"loss": 0.7684,
|
|
"step": 208
|
|
},
|
|
{
|
|
"epoch": 0.69,
|
|
"learning_rate": 5.41304347826087e-06,
|
|
"loss": 0.7686,
|
|
"step": 209
|
|
},
|
|
{
|
|
"epoch": 0.7,
|
|
"learning_rate": 5.4076086956521734e-06,
|
|
"loss": 0.7751,
|
|
"step": 210
|
|
},
|
|
{
|
|
"epoch": 0.7,
|
|
"learning_rate": 5.4021739130434785e-06,
|
|
"loss": 0.7707,
|
|
"step": 211
|
|
},
|
|
{
|
|
"epoch": 0.7,
|
|
"learning_rate": 5.396739130434783e-06,
|
|
"loss": 0.7807,
|
|
"step": 212
|
|
},
|
|
{
|
|
"epoch": 0.71,
|
|
"learning_rate": 5.391304347826087e-06,
|
|
"loss": 0.7735,
|
|
"step": 213
|
|
},
|
|
{
|
|
"epoch": 0.71,
|
|
"learning_rate": 5.385869565217392e-06,
|
|
"loss": 0.7727,
|
|
"step": 214
|
|
},
|
|
{
|
|
"epoch": 0.71,
|
|
"learning_rate": 5.380434782608696e-06,
|
|
"loss": 0.7639,
|
|
"step": 215
|
|
},
|
|
{
|
|
"epoch": 0.72,
|
|
"learning_rate": 5.375e-06,
|
|
"loss": 0.7602,
|
|
"step": 216
|
|
},
|
|
{
|
|
"epoch": 0.72,
|
|
"learning_rate": 5.369565217391304e-06,
|
|
"loss": 0.7832,
|
|
"step": 217
|
|
},
|
|
{
|
|
"epoch": 0.72,
|
|
"learning_rate": 5.364130434782609e-06,
|
|
"loss": 0.7824,
|
|
"step": 218
|
|
},
|
|
{
|
|
"epoch": 0.73,
|
|
"learning_rate": 5.358695652173913e-06,
|
|
"loss": 0.7712,
|
|
"step": 219
|
|
},
|
|
{
|
|
"epoch": 0.73,
|
|
"learning_rate": 5.353260869565218e-06,
|
|
"loss": 0.7614,
|
|
"step": 220
|
|
},
|
|
{
|
|
"epoch": 0.73,
|
|
"learning_rate": 5.347826086956522e-06,
|
|
"loss": 0.782,
|
|
"step": 221
|
|
},
|
|
{
|
|
"epoch": 0.74,
|
|
"learning_rate": 5.342391304347826e-06,
|
|
"loss": 0.7726,
|
|
"step": 222
|
|
},
|
|
{
|
|
"epoch": 0.74,
|
|
"learning_rate": 5.33695652173913e-06,
|
|
"loss": 0.7577,
|
|
"step": 223
|
|
},
|
|
{
|
|
"epoch": 0.74,
|
|
"learning_rate": 5.331521739130435e-06,
|
|
"loss": 0.7623,
|
|
"step": 224
|
|
},
|
|
{
|
|
"epoch": 0.75,
|
|
"learning_rate": 5.326086956521739e-06,
|
|
"loss": 0.761,
|
|
"step": 225
|
|
},
|
|
{
|
|
"epoch": 0.75,
|
|
"learning_rate": 5.320652173913044e-06,
|
|
"loss": 0.782,
|
|
"step": 226
|
|
},
|
|
{
|
|
"epoch": 0.75,
|
|
"learning_rate": 5.315217391304348e-06,
|
|
"loss": 0.768,
|
|
"step": 227
|
|
},
|
|
{
|
|
"epoch": 0.76,
|
|
"learning_rate": 5.309782608695652e-06,
|
|
"loss": 0.7719,
|
|
"step": 228
|
|
},
|
|
{
|
|
"epoch": 0.76,
|
|
"learning_rate": 5.304347826086957e-06,
|
|
"loss": 0.7791,
|
|
"step": 229
|
|
},
|
|
{
|
|
"epoch": 0.76,
|
|
"learning_rate": 5.298913043478261e-06,
|
|
"loss": 0.7725,
|
|
"step": 230
|
|
},
|
|
{
|
|
"epoch": 0.77,
|
|
"learning_rate": 5.2934782608695654e-06,
|
|
"loss": 0.7805,
|
|
"step": 231
|
|
},
|
|
{
|
|
"epoch": 0.77,
|
|
"learning_rate": 5.28804347826087e-06,
|
|
"loss": 0.7689,
|
|
"step": 232
|
|
},
|
|
{
|
|
"epoch": 0.77,
|
|
"learning_rate": 5.282608695652174e-06,
|
|
"loss": 0.7599,
|
|
"step": 233
|
|
},
|
|
{
|
|
"epoch": 0.78,
|
|
"learning_rate": 5.277173913043478e-06,
|
|
"loss": 0.772,
|
|
"step": 234
|
|
},
|
|
{
|
|
"epoch": 0.78,
|
|
"learning_rate": 5.271739130434783e-06,
|
|
"loss": 0.7696,
|
|
"step": 235
|
|
},
|
|
{
|
|
"epoch": 0.78,
|
|
"learning_rate": 5.266304347826087e-06,
|
|
"loss": 0.7774,
|
|
"step": 236
|
|
},
|
|
{
|
|
"epoch": 0.79,
|
|
"learning_rate": 5.260869565217391e-06,
|
|
"loss": 0.7707,
|
|
"step": 237
|
|
},
|
|
{
|
|
"epoch": 0.79,
|
|
"learning_rate": 5.2554347826086955e-06,
|
|
"loss": 0.771,
|
|
"step": 238
|
|
},
|
|
{
|
|
"epoch": 0.79,
|
|
"learning_rate": 5.2500000000000006e-06,
|
|
"loss": 0.7562,
|
|
"step": 239
|
|
},
|
|
{
|
|
"epoch": 0.8,
|
|
"learning_rate": 5.244565217391305e-06,
|
|
"loss": 0.7635,
|
|
"step": 240
|
|
},
|
|
{
|
|
"epoch": 0.8,
|
|
"learning_rate": 5.239130434782609e-06,
|
|
"loss": 0.7594,
|
|
"step": 241
|
|
},
|
|
{
|
|
"epoch": 0.8,
|
|
"learning_rate": 5.233695652173913e-06,
|
|
"loss": 0.7374,
|
|
"step": 242
|
|
},
|
|
{
|
|
"epoch": 0.81,
|
|
"learning_rate": 5.228260869565217e-06,
|
|
"loss": 0.7605,
|
|
"step": 243
|
|
},
|
|
{
|
|
"epoch": 0.81,
|
|
"learning_rate": 5.222826086956522e-06,
|
|
"loss": 0.7742,
|
|
"step": 244
|
|
},
|
|
{
|
|
"epoch": 0.81,
|
|
"eval_loss": 0.7746977806091309,
|
|
"eval_runtime": 17.6216,
|
|
"eval_samples_per_second": 129.5,
|
|
"eval_steps_per_second": 5.448,
|
|
"step": 244
|
|
},
|
|
{
|
|
"epoch": 0.81,
|
|
"learning_rate": 5.2173913043478265e-06,
|
|
"loss": 0.779,
|
|
"step": 245
|
|
},
|
|
{
|
|
"epoch": 0.82,
|
|
"learning_rate": 5.211956521739131e-06,
|
|
"loss": 0.748,
|
|
"step": 246
|
|
},
|
|
{
|
|
"epoch": 0.82,
|
|
"learning_rate": 5.206521739130435e-06,
|
|
"loss": 0.7851,
|
|
"step": 247
|
|
},
|
|
{
|
|
"epoch": 0.82,
|
|
"learning_rate": 5.201086956521739e-06,
|
|
"loss": 0.7543,
|
|
"step": 248
|
|
},
|
|
{
|
|
"epoch": 0.83,
|
|
"learning_rate": 5.195652173913043e-06,
|
|
"loss": 0.7666,
|
|
"step": 249
|
|
},
|
|
{
|
|
"epoch": 0.83,
|
|
"learning_rate": 5.190217391304348e-06,
|
|
"loss": 0.7742,
|
|
"step": 250
|
|
},
|
|
{
|
|
"epoch": 0.83,
|
|
"learning_rate": 5.184782608695652e-06,
|
|
"loss": 0.7726,
|
|
"step": 251
|
|
},
|
|
{
|
|
"epoch": 0.84,
|
|
"learning_rate": 5.1793478260869566e-06,
|
|
"loss": 0.7659,
|
|
"step": 252
|
|
},
|
|
{
|
|
"epoch": 0.84,
|
|
"learning_rate": 5.173913043478262e-06,
|
|
"loss": 0.7677,
|
|
"step": 253
|
|
},
|
|
{
|
|
"epoch": 0.84,
|
|
"learning_rate": 5.168478260869565e-06,
|
|
"loss": 0.7497,
|
|
"step": 254
|
|
},
|
|
{
|
|
"epoch": 0.85,
|
|
"learning_rate": 5.16304347826087e-06,
|
|
"loss": 0.7655,
|
|
"step": 255
|
|
},
|
|
{
|
|
"epoch": 0.85,
|
|
"learning_rate": 5.157608695652174e-06,
|
|
"loss": 0.772,
|
|
"step": 256
|
|
},
|
|
{
|
|
"epoch": 0.85,
|
|
"learning_rate": 5.152173913043478e-06,
|
|
"loss": 0.7765,
|
|
"step": 257
|
|
},
|
|
{
|
|
"epoch": 0.86,
|
|
"learning_rate": 5.1467391304347825e-06,
|
|
"loss": 0.7722,
|
|
"step": 258
|
|
},
|
|
{
|
|
"epoch": 0.86,
|
|
"learning_rate": 5.1413043478260875e-06,
|
|
"loss": 0.7541,
|
|
"step": 259
|
|
},
|
|
{
|
|
"epoch": 0.86,
|
|
"learning_rate": 5.135869565217391e-06,
|
|
"loss": 0.7553,
|
|
"step": 260
|
|
},
|
|
{
|
|
"epoch": 0.87,
|
|
"learning_rate": 5.130434782608696e-06,
|
|
"loss": 0.7713,
|
|
"step": 261
|
|
},
|
|
{
|
|
"epoch": 0.87,
|
|
"learning_rate": 5.125e-06,
|
|
"loss": 0.7649,
|
|
"step": 262
|
|
},
|
|
{
|
|
"epoch": 0.87,
|
|
"learning_rate": 5.119565217391304e-06,
|
|
"loss": 0.7639,
|
|
"step": 263
|
|
},
|
|
{
|
|
"epoch": 0.88,
|
|
"learning_rate": 5.114130434782608e-06,
|
|
"loss": 0.786,
|
|
"step": 264
|
|
},
|
|
{
|
|
"epoch": 0.88,
|
|
"learning_rate": 5.1086956521739134e-06,
|
|
"loss": 0.7542,
|
|
"step": 265
|
|
},
|
|
{
|
|
"epoch": 0.88,
|
|
"learning_rate": 5.103260869565218e-06,
|
|
"loss": 0.765,
|
|
"step": 266
|
|
},
|
|
{
|
|
"epoch": 0.89,
|
|
"learning_rate": 5.097826086956522e-06,
|
|
"loss": 0.7495,
|
|
"step": 267
|
|
},
|
|
{
|
|
"epoch": 0.89,
|
|
"learning_rate": 5.092391304347827e-06,
|
|
"loss": 0.7719,
|
|
"step": 268
|
|
},
|
|
{
|
|
"epoch": 0.89,
|
|
"learning_rate": 5.08695652173913e-06,
|
|
"loss": 0.774,
|
|
"step": 269
|
|
},
|
|
{
|
|
"epoch": 0.9,
|
|
"learning_rate": 5.081521739130435e-06,
|
|
"loss": 0.7665,
|
|
"step": 270
|
|
},
|
|
{
|
|
"epoch": 0.9,
|
|
"learning_rate": 5.076086956521739e-06,
|
|
"loss": 0.7672,
|
|
"step": 271
|
|
},
|
|
{
|
|
"epoch": 0.9,
|
|
"learning_rate": 5.0706521739130435e-06,
|
|
"loss": 0.7601,
|
|
"step": 272
|
|
},
|
|
{
|
|
"epoch": 0.91,
|
|
"learning_rate": 5.065217391304348e-06,
|
|
"loss": 0.7602,
|
|
"step": 273
|
|
},
|
|
{
|
|
"epoch": 0.91,
|
|
"learning_rate": 5.059782608695653e-06,
|
|
"loss": 0.753,
|
|
"step": 274
|
|
},
|
|
{
|
|
"epoch": 0.91,
|
|
"learning_rate": 5.054347826086956e-06,
|
|
"loss": 0.7421,
|
|
"step": 275
|
|
},
|
|
{
|
|
"epoch": 0.92,
|
|
"learning_rate": 5.048913043478261e-06,
|
|
"loss": 0.7616,
|
|
"step": 276
|
|
},
|
|
{
|
|
"epoch": 0.92,
|
|
"learning_rate": 5.043478260869565e-06,
|
|
"loss": 0.7638,
|
|
"step": 277
|
|
},
|
|
{
|
|
"epoch": 0.92,
|
|
"learning_rate": 5.0380434782608695e-06,
|
|
"loss": 0.7704,
|
|
"step": 278
|
|
},
|
|
{
|
|
"epoch": 0.93,
|
|
"learning_rate": 5.032608695652174e-06,
|
|
"loss": 0.7513,
|
|
"step": 279
|
|
},
|
|
{
|
|
"epoch": 0.93,
|
|
"learning_rate": 5.027173913043479e-06,
|
|
"loss": 0.7916,
|
|
"step": 280
|
|
},
|
|
{
|
|
"epoch": 0.93,
|
|
"learning_rate": 5.021739130434783e-06,
|
|
"loss": 0.7544,
|
|
"step": 281
|
|
},
|
|
{
|
|
"epoch": 0.94,
|
|
"learning_rate": 5.016304347826087e-06,
|
|
"loss": 0.7646,
|
|
"step": 282
|
|
},
|
|
{
|
|
"epoch": 0.94,
|
|
"learning_rate": 5.010869565217392e-06,
|
|
"loss": 0.7663,
|
|
"step": 283
|
|
},
|
|
{
|
|
"epoch": 0.94,
|
|
"learning_rate": 5.005434782608695e-06,
|
|
"loss": 0.7605,
|
|
"step": 284
|
|
},
|
|
{
|
|
"epoch": 0.95,
|
|
"learning_rate": 5e-06,
|
|
"loss": 0.7634,
|
|
"step": 285
|
|
},
|
|
{
|
|
"epoch": 0.95,
|
|
"learning_rate": 4.994565217391305e-06,
|
|
"loss": 0.7702,
|
|
"step": 286
|
|
},
|
|
{
|
|
"epoch": 0.95,
|
|
"learning_rate": 4.989130434782609e-06,
|
|
"loss": 0.7531,
|
|
"step": 287
|
|
},
|
|
{
|
|
"epoch": 0.96,
|
|
"learning_rate": 4.983695652173913e-06,
|
|
"loss": 0.7609,
|
|
"step": 288
|
|
},
|
|
{
|
|
"epoch": 0.96,
|
|
"learning_rate": 4.978260869565218e-06,
|
|
"loss": 0.7697,
|
|
"step": 289
|
|
},
|
|
{
|
|
"epoch": 0.96,
|
|
"learning_rate": 4.972826086956521e-06,
|
|
"loss": 0.7531,
|
|
"step": 290
|
|
},
|
|
{
|
|
"epoch": 0.97,
|
|
"learning_rate": 4.967391304347826e-06,
|
|
"loss": 0.7438,
|
|
"step": 291
|
|
},
|
|
{
|
|
"epoch": 0.97,
|
|
"learning_rate": 4.9619565217391305e-06,
|
|
"loss": 0.7783,
|
|
"step": 292
|
|
},
|
|
{
|
|
"epoch": 0.97,
|
|
"learning_rate": 4.956521739130435e-06,
|
|
"loss": 0.7668,
|
|
"step": 293
|
|
},
|
|
{
|
|
"epoch": 0.98,
|
|
"learning_rate": 4.951086956521739e-06,
|
|
"loss": 0.8017,
|
|
"step": 294
|
|
},
|
|
{
|
|
"epoch": 0.98,
|
|
"learning_rate": 4.945652173913044e-06,
|
|
"loss": 0.7361,
|
|
"step": 295
|
|
},
|
|
{
|
|
"epoch": 0.98,
|
|
"learning_rate": 4.940217391304348e-06,
|
|
"loss": 0.7649,
|
|
"step": 296
|
|
},
|
|
{
|
|
"epoch": 0.99,
|
|
"learning_rate": 4.934782608695652e-06,
|
|
"loss": 0.7704,
|
|
"step": 297
|
|
},
|
|
{
|
|
"epoch": 0.99,
|
|
"learning_rate": 4.929347826086957e-06,
|
|
"loss": 0.7502,
|
|
"step": 298
|
|
},
|
|
{
|
|
"epoch": 0.99,
|
|
"learning_rate": 4.923913043478261e-06,
|
|
"loss": 0.7736,
|
|
"step": 299
|
|
},
|
|
{
|
|
"epoch": 1.0,
|
|
"learning_rate": 4.918478260869566e-06,
|
|
"loss": 0.759,
|
|
"step": 300
|
|
},
|
|
{
|
|
"epoch": 1.0,
|
|
"learning_rate": 4.91304347826087e-06,
|
|
"loss": 0.7812,
|
|
"step": 301
|
|
},
|
|
{
|
|
"epoch": 1.0,
|
|
"learning_rate": 4.907608695652174e-06,
|
|
"loss": 0.7343,
|
|
"step": 302
|
|
},
|
|
{
|
|
"epoch": 1.0,
|
|
"learning_rate": 4.902173913043478e-06,
|
|
"loss": 0.7192,
|
|
"step": 303
|
|
},
|
|
{
|
|
"epoch": 1.01,
|
|
"learning_rate": 4.896739130434783e-06,
|
|
"loss": 0.7057,
|
|
"step": 304
|
|
},
|
|
{
|
|
"epoch": 1.01,
|
|
"learning_rate": 4.8913043478260865e-06,
|
|
"loss": 0.6932,
|
|
"step": 305
|
|
},
|
|
{
|
|
"epoch": 1.01,
|
|
"eval_loss": 0.7692939639091492,
|
|
"eval_runtime": 17.6192,
|
|
"eval_samples_per_second": 129.518,
|
|
"eval_steps_per_second": 5.449,
|
|
"step": 305
|
|
},
|
|
{
|
|
"epoch": 1.01,
|
|
"learning_rate": 4.8858695652173916e-06,
|
|
"loss": 0.711,
|
|
"step": 306
|
|
},
|
|
{
|
|
"epoch": 1.02,
|
|
"learning_rate": 4.880434782608696e-06,
|
|
"loss": 0.7022,
|
|
"step": 307
|
|
},
|
|
{
|
|
"epoch": 1.02,
|
|
"learning_rate": 4.875e-06,
|
|
"loss": 0.7019,
|
|
"step": 308
|
|
},
|
|
{
|
|
"epoch": 1.02,
|
|
"learning_rate": 4.869565217391305e-06,
|
|
"loss": 0.7355,
|
|
"step": 309
|
|
},
|
|
{
|
|
"epoch": 1.03,
|
|
"learning_rate": 4.864130434782609e-06,
|
|
"loss": 0.6934,
|
|
"step": 310
|
|
},
|
|
{
|
|
"epoch": 1.03,
|
|
"learning_rate": 4.858695652173913e-06,
|
|
"loss": 0.7088,
|
|
"step": 311
|
|
},
|
|
{
|
|
"epoch": 1.03,
|
|
"learning_rate": 4.8532608695652175e-06,
|
|
"loss": 0.7062,
|
|
"step": 312
|
|
},
|
|
{
|
|
"epoch": 1.04,
|
|
"learning_rate": 4.847826086956522e-06,
|
|
"loss": 0.6968,
|
|
"step": 313
|
|
},
|
|
{
|
|
"epoch": 1.04,
|
|
"learning_rate": 4.842391304347826e-06,
|
|
"loss": 0.7053,
|
|
"step": 314
|
|
},
|
|
{
|
|
"epoch": 1.04,
|
|
"learning_rate": 4.836956521739131e-06,
|
|
"loss": 0.7199,
|
|
"step": 315
|
|
},
|
|
{
|
|
"epoch": 1.05,
|
|
"learning_rate": 4.831521739130435e-06,
|
|
"loss": 0.6892,
|
|
"step": 316
|
|
},
|
|
{
|
|
"epoch": 1.05,
|
|
"learning_rate": 4.826086956521739e-06,
|
|
"loss": 0.6913,
|
|
"step": 317
|
|
},
|
|
{
|
|
"epoch": 1.05,
|
|
"learning_rate": 4.820652173913043e-06,
|
|
"loss": 0.6999,
|
|
"step": 318
|
|
},
|
|
{
|
|
"epoch": 1.06,
|
|
"learning_rate": 4.815217391304348e-06,
|
|
"loss": 0.7066,
|
|
"step": 319
|
|
},
|
|
{
|
|
"epoch": 1.06,
|
|
"learning_rate": 4.809782608695652e-06,
|
|
"loss": 0.7163,
|
|
"step": 320
|
|
},
|
|
{
|
|
"epoch": 1.06,
|
|
"learning_rate": 4.804347826086957e-06,
|
|
"loss": 0.7102,
|
|
"step": 321
|
|
},
|
|
{
|
|
"epoch": 1.07,
|
|
"learning_rate": 4.798913043478261e-06,
|
|
"loss": 0.7012,
|
|
"step": 322
|
|
},
|
|
{
|
|
"epoch": 1.07,
|
|
"learning_rate": 4.793478260869565e-06,
|
|
"loss": 0.6949,
|
|
"step": 323
|
|
},
|
|
{
|
|
"epoch": 1.07,
|
|
"learning_rate": 4.78804347826087e-06,
|
|
"loss": 0.6825,
|
|
"step": 324
|
|
},
|
|
{
|
|
"epoch": 1.08,
|
|
"learning_rate": 4.782608695652174e-06,
|
|
"loss": 0.6905,
|
|
"step": 325
|
|
},
|
|
{
|
|
"epoch": 1.08,
|
|
"learning_rate": 4.7771739130434785e-06,
|
|
"loss": 0.709,
|
|
"step": 326
|
|
},
|
|
{
|
|
"epoch": 1.08,
|
|
"learning_rate": 4.771739130434783e-06,
|
|
"loss": 0.6833,
|
|
"step": 327
|
|
},
|
|
{
|
|
"epoch": 1.09,
|
|
"learning_rate": 4.766304347826087e-06,
|
|
"loss": 0.6943,
|
|
"step": 328
|
|
},
|
|
{
|
|
"epoch": 1.09,
|
|
"learning_rate": 4.760869565217391e-06,
|
|
"loss": 0.701,
|
|
"step": 329
|
|
},
|
|
{
|
|
"epoch": 1.09,
|
|
"learning_rate": 4.755434782608696e-06,
|
|
"loss": 0.6976,
|
|
"step": 330
|
|
},
|
|
{
|
|
"epoch": 1.1,
|
|
"learning_rate": 4.75e-06,
|
|
"loss": 0.6999,
|
|
"step": 331
|
|
},
|
|
{
|
|
"epoch": 1.1,
|
|
"learning_rate": 4.7445652173913044e-06,
|
|
"loss": 0.708,
|
|
"step": 332
|
|
},
|
|
{
|
|
"epoch": 1.1,
|
|
"learning_rate": 4.739130434782609e-06,
|
|
"loss": 0.7215,
|
|
"step": 333
|
|
},
|
|
{
|
|
"epoch": 1.11,
|
|
"learning_rate": 4.733695652173913e-06,
|
|
"loss": 0.6919,
|
|
"step": 334
|
|
},
|
|
{
|
|
"epoch": 1.11,
|
|
"learning_rate": 4.728260869565217e-06,
|
|
"loss": 0.7093,
|
|
"step": 335
|
|
},
|
|
{
|
|
"epoch": 1.11,
|
|
"learning_rate": 4.722826086956522e-06,
|
|
"loss": 0.6882,
|
|
"step": 336
|
|
},
|
|
{
|
|
"epoch": 1.12,
|
|
"learning_rate": 4.717391304347826e-06,
|
|
"loss": 0.7107,
|
|
"step": 337
|
|
},
|
|
{
|
|
"epoch": 1.12,
|
|
"learning_rate": 4.71195652173913e-06,
|
|
"loss": 0.692,
|
|
"step": 338
|
|
},
|
|
{
|
|
"epoch": 1.12,
|
|
"learning_rate": 4.706521739130435e-06,
|
|
"loss": 0.7042,
|
|
"step": 339
|
|
},
|
|
{
|
|
"epoch": 1.13,
|
|
"learning_rate": 4.7010869565217396e-06,
|
|
"loss": 0.7004,
|
|
"step": 340
|
|
},
|
|
{
|
|
"epoch": 1.13,
|
|
"learning_rate": 4.695652173913044e-06,
|
|
"loss": 0.7049,
|
|
"step": 341
|
|
},
|
|
{
|
|
"epoch": 1.13,
|
|
"learning_rate": 4.690217391304348e-06,
|
|
"loss": 0.7071,
|
|
"step": 342
|
|
},
|
|
{
|
|
"epoch": 1.14,
|
|
"learning_rate": 4.684782608695652e-06,
|
|
"loss": 0.6917,
|
|
"step": 343
|
|
},
|
|
{
|
|
"epoch": 1.14,
|
|
"learning_rate": 4.679347826086956e-06,
|
|
"loss": 0.7141,
|
|
"step": 344
|
|
},
|
|
{
|
|
"epoch": 1.14,
|
|
"learning_rate": 4.673913043478261e-06,
|
|
"loss": 0.7129,
|
|
"step": 345
|
|
},
|
|
{
|
|
"epoch": 1.15,
|
|
"learning_rate": 4.6684782608695655e-06,
|
|
"loss": 0.6929,
|
|
"step": 346
|
|
},
|
|
{
|
|
"epoch": 1.15,
|
|
"learning_rate": 4.66304347826087e-06,
|
|
"loss": 0.7041,
|
|
"step": 347
|
|
},
|
|
{
|
|
"epoch": 1.15,
|
|
"learning_rate": 4.657608695652174e-06,
|
|
"loss": 0.6764,
|
|
"step": 348
|
|
},
|
|
{
|
|
"epoch": 1.16,
|
|
"learning_rate": 4.652173913043478e-06,
|
|
"loss": 0.6999,
|
|
"step": 349
|
|
},
|
|
{
|
|
"epoch": 1.16,
|
|
"learning_rate": 4.646739130434783e-06,
|
|
"loss": 0.7008,
|
|
"step": 350
|
|
},
|
|
{
|
|
"epoch": 1.16,
|
|
"learning_rate": 4.641304347826087e-06,
|
|
"loss": 0.7005,
|
|
"step": 351
|
|
},
|
|
{
|
|
"epoch": 1.17,
|
|
"learning_rate": 4.635869565217391e-06,
|
|
"loss": 0.7069,
|
|
"step": 352
|
|
},
|
|
{
|
|
"epoch": 1.17,
|
|
"learning_rate": 4.630434782608696e-06,
|
|
"loss": 0.6915,
|
|
"step": 353
|
|
},
|
|
{
|
|
"epoch": 1.17,
|
|
"learning_rate": 4.625000000000001e-06,
|
|
"loss": 0.6864,
|
|
"step": 354
|
|
},
|
|
{
|
|
"epoch": 1.18,
|
|
"learning_rate": 4.619565217391304e-06,
|
|
"loss": 0.6933,
|
|
"step": 355
|
|
},
|
|
{
|
|
"epoch": 1.18,
|
|
"learning_rate": 4.614130434782609e-06,
|
|
"loss": 0.6874,
|
|
"step": 356
|
|
},
|
|
{
|
|
"epoch": 1.18,
|
|
"learning_rate": 4.608695652173913e-06,
|
|
"loss": 0.695,
|
|
"step": 357
|
|
},
|
|
{
|
|
"epoch": 1.19,
|
|
"learning_rate": 4.603260869565217e-06,
|
|
"loss": 0.6937,
|
|
"step": 358
|
|
},
|
|
{
|
|
"epoch": 1.19,
|
|
"learning_rate": 4.5978260869565215e-06,
|
|
"loss": 0.714,
|
|
"step": 359
|
|
},
|
|
{
|
|
"epoch": 1.19,
|
|
"learning_rate": 4.5923913043478265e-06,
|
|
"loss": 0.7005,
|
|
"step": 360
|
|
},
|
|
{
|
|
"epoch": 1.2,
|
|
"learning_rate": 4.58695652173913e-06,
|
|
"loss": 0.701,
|
|
"step": 361
|
|
},
|
|
{
|
|
"epoch": 1.2,
|
|
"learning_rate": 4.581521739130435e-06,
|
|
"loss": 0.7004,
|
|
"step": 362
|
|
},
|
|
{
|
|
"epoch": 1.2,
|
|
"learning_rate": 4.576086956521739e-06,
|
|
"loss": 0.7059,
|
|
"step": 363
|
|
},
|
|
{
|
|
"epoch": 1.21,
|
|
"learning_rate": 4.570652173913043e-06,
|
|
"loss": 0.7226,
|
|
"step": 364
|
|
},
|
|
{
|
|
"epoch": 1.21,
|
|
"learning_rate": 4.565217391304348e-06,
|
|
"loss": 0.7113,
|
|
"step": 365
|
|
},
|
|
{
|
|
"epoch": 1.21,
|
|
"learning_rate": 4.5597826086956525e-06,
|
|
"loss": 0.6889,
|
|
"step": 366
|
|
},
|
|
{
|
|
"epoch": 1.21,
|
|
"eval_loss": 0.7673465609550476,
|
|
"eval_runtime": 17.6178,
|
|
"eval_samples_per_second": 129.528,
|
|
"eval_steps_per_second": 5.449,
|
|
"step": 366
|
|
},
|
|
{
|
|
"epoch": 1.22,
|
|
"learning_rate": 4.554347826086957e-06,
|
|
"loss": 0.7108,
|
|
"step": 367
|
|
},
|
|
{
|
|
"epoch": 1.22,
|
|
"learning_rate": 4.548913043478261e-06,
|
|
"loss": 0.677,
|
|
"step": 368
|
|
},
|
|
{
|
|
"epoch": 1.22,
|
|
"learning_rate": 4.543478260869566e-06,
|
|
"loss": 0.6995,
|
|
"step": 369
|
|
},
|
|
{
|
|
"epoch": 1.23,
|
|
"learning_rate": 4.538043478260869e-06,
|
|
"loss": 0.6886,
|
|
"step": 370
|
|
},
|
|
{
|
|
"epoch": 1.23,
|
|
"learning_rate": 4.532608695652174e-06,
|
|
"loss": 0.7069,
|
|
"step": 371
|
|
},
|
|
{
|
|
"epoch": 1.23,
|
|
"learning_rate": 4.527173913043478e-06,
|
|
"loss": 0.6865,
|
|
"step": 372
|
|
},
|
|
{
|
|
"epoch": 1.24,
|
|
"learning_rate": 4.5217391304347826e-06,
|
|
"loss": 0.7159,
|
|
"step": 373
|
|
},
|
|
{
|
|
"epoch": 1.24,
|
|
"learning_rate": 4.516304347826087e-06,
|
|
"loss": 0.7085,
|
|
"step": 374
|
|
},
|
|
{
|
|
"epoch": 1.24,
|
|
"learning_rate": 4.510869565217392e-06,
|
|
"loss": 0.7113,
|
|
"step": 375
|
|
},
|
|
{
|
|
"epoch": 1.25,
|
|
"learning_rate": 4.505434782608695e-06,
|
|
"loss": 0.7031,
|
|
"step": 376
|
|
},
|
|
{
|
|
"epoch": 1.25,
|
|
"learning_rate": 4.5e-06,
|
|
"loss": 0.7236,
|
|
"step": 377
|
|
},
|
|
{
|
|
"epoch": 1.25,
|
|
"learning_rate": 4.494565217391305e-06,
|
|
"loss": 0.6915,
|
|
"step": 378
|
|
},
|
|
{
|
|
"epoch": 1.26,
|
|
"learning_rate": 4.4891304347826085e-06,
|
|
"loss": 0.7112,
|
|
"step": 379
|
|
},
|
|
{
|
|
"epoch": 1.26,
|
|
"learning_rate": 4.4836956521739135e-06,
|
|
"loss": 0.6947,
|
|
"step": 380
|
|
},
|
|
{
|
|
"epoch": 1.26,
|
|
"learning_rate": 4.478260869565218e-06,
|
|
"loss": 0.7108,
|
|
"step": 381
|
|
},
|
|
{
|
|
"epoch": 1.27,
|
|
"learning_rate": 4.472826086956522e-06,
|
|
"loss": 0.7302,
|
|
"step": 382
|
|
},
|
|
{
|
|
"epoch": 1.27,
|
|
"learning_rate": 4.467391304347826e-06,
|
|
"loss": 0.7203,
|
|
"step": 383
|
|
},
|
|
{
|
|
"epoch": 1.27,
|
|
"learning_rate": 4.461956521739131e-06,
|
|
"loss": 0.6911,
|
|
"step": 384
|
|
},
|
|
{
|
|
"epoch": 1.28,
|
|
"learning_rate": 4.456521739130434e-06,
|
|
"loss": 0.7072,
|
|
"step": 385
|
|
},
|
|
{
|
|
"epoch": 1.28,
|
|
"learning_rate": 4.451086956521739e-06,
|
|
"loss": 0.698,
|
|
"step": 386
|
|
},
|
|
{
|
|
"epoch": 1.28,
|
|
"learning_rate": 4.445652173913044e-06,
|
|
"loss": 0.7089,
|
|
"step": 387
|
|
},
|
|
{
|
|
"epoch": 1.29,
|
|
"learning_rate": 4.440217391304348e-06,
|
|
"loss": 0.7027,
|
|
"step": 388
|
|
},
|
|
{
|
|
"epoch": 1.29,
|
|
"learning_rate": 4.434782608695652e-06,
|
|
"loss": 0.6814,
|
|
"step": 389
|
|
},
|
|
{
|
|
"epoch": 1.29,
|
|
"learning_rate": 4.429347826086957e-06,
|
|
"loss": 0.7034,
|
|
"step": 390
|
|
},
|
|
{
|
|
"epoch": 1.3,
|
|
"learning_rate": 4.423913043478261e-06,
|
|
"loss": 0.6952,
|
|
"step": 391
|
|
},
|
|
{
|
|
"epoch": 1.3,
|
|
"learning_rate": 4.418478260869565e-06,
|
|
"loss": 0.7081,
|
|
"step": 392
|
|
},
|
|
{
|
|
"epoch": 1.3,
|
|
"learning_rate": 4.41304347826087e-06,
|
|
"loss": 0.711,
|
|
"step": 393
|
|
},
|
|
{
|
|
"epoch": 1.31,
|
|
"learning_rate": 4.407608695652174e-06,
|
|
"loss": 0.7107,
|
|
"step": 394
|
|
},
|
|
{
|
|
"epoch": 1.31,
|
|
"learning_rate": 4.402173913043479e-06,
|
|
"loss": 0.6898,
|
|
"step": 395
|
|
},
|
|
{
|
|
"epoch": 1.31,
|
|
"learning_rate": 4.396739130434783e-06,
|
|
"loss": 0.712,
|
|
"step": 396
|
|
},
|
|
{
|
|
"epoch": 1.32,
|
|
"learning_rate": 4.391304347826087e-06,
|
|
"loss": 0.6985,
|
|
"step": 397
|
|
},
|
|
{
|
|
"epoch": 1.32,
|
|
"learning_rate": 4.385869565217391e-06,
|
|
"loss": 0.7024,
|
|
"step": 398
|
|
},
|
|
{
|
|
"epoch": 1.32,
|
|
"learning_rate": 4.380434782608696e-06,
|
|
"loss": 0.7142,
|
|
"step": 399
|
|
},
|
|
{
|
|
"epoch": 1.33,
|
|
"learning_rate": 4.375e-06,
|
|
"loss": 0.7041,
|
|
"step": 400
|
|
},
|
|
{
|
|
"epoch": 1.33,
|
|
"learning_rate": 4.369565217391305e-06,
|
|
"loss": 0.6963,
|
|
"step": 401
|
|
},
|
|
{
|
|
"epoch": 1.33,
|
|
"learning_rate": 4.364130434782609e-06,
|
|
"loss": 0.7048,
|
|
"step": 402
|
|
},
|
|
{
|
|
"epoch": 1.34,
|
|
"learning_rate": 4.358695652173913e-06,
|
|
"loss": 0.6865,
|
|
"step": 403
|
|
},
|
|
{
|
|
"epoch": 1.34,
|
|
"learning_rate": 4.353260869565217e-06,
|
|
"loss": 0.721,
|
|
"step": 404
|
|
},
|
|
{
|
|
"epoch": 1.34,
|
|
"learning_rate": 4.347826086956522e-06,
|
|
"loss": 0.6925,
|
|
"step": 405
|
|
},
|
|
{
|
|
"epoch": 1.35,
|
|
"learning_rate": 4.342391304347826e-06,
|
|
"loss": 0.7016,
|
|
"step": 406
|
|
},
|
|
{
|
|
"epoch": 1.35,
|
|
"learning_rate": 4.3369565217391306e-06,
|
|
"loss": 0.7164,
|
|
"step": 407
|
|
},
|
|
{
|
|
"epoch": 1.35,
|
|
"learning_rate": 4.331521739130435e-06,
|
|
"loss": 0.6937,
|
|
"step": 408
|
|
},
|
|
{
|
|
"epoch": 1.36,
|
|
"learning_rate": 4.326086956521739e-06,
|
|
"loss": 0.7193,
|
|
"step": 409
|
|
},
|
|
{
|
|
"epoch": 1.36,
|
|
"learning_rate": 4.320652173913044e-06,
|
|
"loss": 0.6959,
|
|
"step": 410
|
|
},
|
|
{
|
|
"epoch": 1.36,
|
|
"learning_rate": 4.315217391304348e-06,
|
|
"loss": 0.6899,
|
|
"step": 411
|
|
},
|
|
{
|
|
"epoch": 1.37,
|
|
"learning_rate": 4.309782608695652e-06,
|
|
"loss": 0.7052,
|
|
"step": 412
|
|
},
|
|
{
|
|
"epoch": 1.37,
|
|
"learning_rate": 4.3043478260869565e-06,
|
|
"loss": 0.7075,
|
|
"step": 413
|
|
},
|
|
{
|
|
"epoch": 1.37,
|
|
"learning_rate": 4.298913043478261e-06,
|
|
"loss": 0.7107,
|
|
"step": 414
|
|
},
|
|
{
|
|
"epoch": 1.38,
|
|
"learning_rate": 4.293478260869565e-06,
|
|
"loss": 0.7055,
|
|
"step": 415
|
|
},
|
|
{
|
|
"epoch": 1.38,
|
|
"learning_rate": 4.28804347826087e-06,
|
|
"loss": 0.7116,
|
|
"step": 416
|
|
},
|
|
{
|
|
"epoch": 1.38,
|
|
"learning_rate": 4.282608695652174e-06,
|
|
"loss": 0.7202,
|
|
"step": 417
|
|
},
|
|
{
|
|
"epoch": 1.39,
|
|
"learning_rate": 4.277173913043478e-06,
|
|
"loss": 0.6763,
|
|
"step": 418
|
|
},
|
|
{
|
|
"epoch": 1.39,
|
|
"learning_rate": 4.271739130434783e-06,
|
|
"loss": 0.6957,
|
|
"step": 419
|
|
},
|
|
{
|
|
"epoch": 1.39,
|
|
"learning_rate": 4.2663043478260874e-06,
|
|
"loss": 0.7022,
|
|
"step": 420
|
|
},
|
|
{
|
|
"epoch": 1.4,
|
|
"learning_rate": 4.260869565217392e-06,
|
|
"loss": 0.7079,
|
|
"step": 421
|
|
},
|
|
{
|
|
"epoch": 1.4,
|
|
"learning_rate": 4.255434782608696e-06,
|
|
"loss": 0.7023,
|
|
"step": 422
|
|
},
|
|
{
|
|
"epoch": 1.4,
|
|
"learning_rate": 4.25e-06,
|
|
"loss": 0.6954,
|
|
"step": 423
|
|
},
|
|
{
|
|
"epoch": 1.41,
|
|
"learning_rate": 4.244565217391304e-06,
|
|
"loss": 0.693,
|
|
"step": 424
|
|
},
|
|
{
|
|
"epoch": 1.41,
|
|
"learning_rate": 4.239130434782609e-06,
|
|
"loss": 0.7061,
|
|
"step": 425
|
|
},
|
|
{
|
|
"epoch": 1.41,
|
|
"learning_rate": 4.233695652173913e-06,
|
|
"loss": 0.701,
|
|
"step": 426
|
|
},
|
|
{
|
|
"epoch": 1.42,
|
|
"learning_rate": 4.2282608695652175e-06,
|
|
"loss": 0.708,
|
|
"step": 427
|
|
},
|
|
{
|
|
"epoch": 1.42,
|
|
"eval_loss": 0.7639372944831848,
|
|
"eval_runtime": 17.6159,
|
|
"eval_samples_per_second": 129.542,
|
|
"eval_steps_per_second": 5.45,
|
|
"step": 427
|
|
},
|
|
{
|
|
"epoch": 1.42,
|
|
"learning_rate": 4.222826086956522e-06,
|
|
"loss": 0.6991,
|
|
"step": 428
|
|
},
|
|
{
|
|
"epoch": 1.42,
|
|
"learning_rate": 4.217391304347826e-06,
|
|
"loss": 0.7042,
|
|
"step": 429
|
|
},
|
|
{
|
|
"epoch": 1.43,
|
|
"learning_rate": 4.21195652173913e-06,
|
|
"loss": 0.6984,
|
|
"step": 430
|
|
},
|
|
{
|
|
"epoch": 1.43,
|
|
"learning_rate": 4.206521739130435e-06,
|
|
"loss": 0.6925,
|
|
"step": 431
|
|
},
|
|
{
|
|
"epoch": 1.43,
|
|
"learning_rate": 4.201086956521739e-06,
|
|
"loss": 0.6928,
|
|
"step": 432
|
|
},
|
|
{
|
|
"epoch": 1.44,
|
|
"learning_rate": 4.1956521739130434e-06,
|
|
"loss": 0.7,
|
|
"step": 433
|
|
},
|
|
{
|
|
"epoch": 1.44,
|
|
"learning_rate": 4.1902173913043485e-06,
|
|
"loss": 0.6964,
|
|
"step": 434
|
|
},
|
|
{
|
|
"epoch": 1.44,
|
|
"learning_rate": 4.184782608695652e-06,
|
|
"loss": 0.6815,
|
|
"step": 435
|
|
},
|
|
{
|
|
"epoch": 1.45,
|
|
"learning_rate": 4.179347826086957e-06,
|
|
"loss": 0.7096,
|
|
"step": 436
|
|
},
|
|
{
|
|
"epoch": 1.45,
|
|
"learning_rate": 4.173913043478261e-06,
|
|
"loss": 0.6967,
|
|
"step": 437
|
|
},
|
|
{
|
|
"epoch": 1.45,
|
|
"learning_rate": 4.168478260869565e-06,
|
|
"loss": 0.6979,
|
|
"step": 438
|
|
},
|
|
{
|
|
"epoch": 1.46,
|
|
"learning_rate": 4.163043478260869e-06,
|
|
"loss": 0.7085,
|
|
"step": 439
|
|
},
|
|
{
|
|
"epoch": 1.46,
|
|
"learning_rate": 4.157608695652174e-06,
|
|
"loss": 0.7059,
|
|
"step": 440
|
|
},
|
|
{
|
|
"epoch": 1.46,
|
|
"learning_rate": 4.1521739130434786e-06,
|
|
"loss": 0.6915,
|
|
"step": 441
|
|
},
|
|
{
|
|
"epoch": 1.47,
|
|
"learning_rate": 4.146739130434783e-06,
|
|
"loss": 0.6905,
|
|
"step": 442
|
|
},
|
|
{
|
|
"epoch": 1.47,
|
|
"learning_rate": 4.141304347826087e-06,
|
|
"loss": 0.7023,
|
|
"step": 443
|
|
},
|
|
{
|
|
"epoch": 1.47,
|
|
"learning_rate": 4.135869565217391e-06,
|
|
"loss": 0.7031,
|
|
"step": 444
|
|
},
|
|
{
|
|
"epoch": 1.48,
|
|
"learning_rate": 4.130434782608695e-06,
|
|
"loss": 0.6959,
|
|
"step": 445
|
|
},
|
|
{
|
|
"epoch": 1.48,
|
|
"learning_rate": 4.125e-06,
|
|
"loss": 0.7103,
|
|
"step": 446
|
|
},
|
|
{
|
|
"epoch": 1.48,
|
|
"learning_rate": 4.1195652173913045e-06,
|
|
"loss": 0.7052,
|
|
"step": 447
|
|
},
|
|
{
|
|
"epoch": 1.49,
|
|
"learning_rate": 4.114130434782609e-06,
|
|
"loss": 0.6919,
|
|
"step": 448
|
|
},
|
|
{
|
|
"epoch": 1.49,
|
|
"learning_rate": 4.108695652173914e-06,
|
|
"loss": 0.6965,
|
|
"step": 449
|
|
},
|
|
{
|
|
"epoch": 1.49,
|
|
"learning_rate": 4.103260869565217e-06,
|
|
"loss": 0.6912,
|
|
"step": 450
|
|
},
|
|
{
|
|
"epoch": 1.5,
|
|
"learning_rate": 4.097826086956522e-06,
|
|
"loss": 0.6924,
|
|
"step": 451
|
|
},
|
|
{
|
|
"epoch": 1.5,
|
|
"learning_rate": 4.092391304347826e-06,
|
|
"loss": 0.6991,
|
|
"step": 452
|
|
},
|
|
{
|
|
"epoch": 1.5,
|
|
"learning_rate": 4.08695652173913e-06,
|
|
"loss": 0.7091,
|
|
"step": 453
|
|
},
|
|
{
|
|
"epoch": 1.51,
|
|
"learning_rate": 4.081521739130435e-06,
|
|
"loss": 0.6905,
|
|
"step": 454
|
|
},
|
|
{
|
|
"epoch": 1.51,
|
|
"learning_rate": 4.07608695652174e-06,
|
|
"loss": 0.6723,
|
|
"step": 455
|
|
},
|
|
{
|
|
"epoch": 1.51,
|
|
"learning_rate": 4.070652173913043e-06,
|
|
"loss": 0.7039,
|
|
"step": 456
|
|
},
|
|
{
|
|
"epoch": 1.52,
|
|
"learning_rate": 4.065217391304348e-06,
|
|
"loss": 0.7229,
|
|
"step": 457
|
|
},
|
|
{
|
|
"epoch": 1.52,
|
|
"learning_rate": 4.059782608695652e-06,
|
|
"loss": 0.7003,
|
|
"step": 458
|
|
},
|
|
{
|
|
"epoch": 1.52,
|
|
"learning_rate": 4.054347826086956e-06,
|
|
"loss": 0.7108,
|
|
"step": 459
|
|
},
|
|
{
|
|
"epoch": 1.53,
|
|
"learning_rate": 4.048913043478261e-06,
|
|
"loss": 0.7016,
|
|
"step": 460
|
|
},
|
|
{
|
|
"epoch": 1.53,
|
|
"learning_rate": 4.0434782608695655e-06,
|
|
"loss": 0.6989,
|
|
"step": 461
|
|
},
|
|
{
|
|
"epoch": 1.53,
|
|
"learning_rate": 4.03804347826087e-06,
|
|
"loss": 0.6961,
|
|
"step": 462
|
|
},
|
|
{
|
|
"epoch": 1.54,
|
|
"learning_rate": 4.032608695652174e-06,
|
|
"loss": 0.7032,
|
|
"step": 463
|
|
},
|
|
{
|
|
"epoch": 1.54,
|
|
"learning_rate": 4.027173913043479e-06,
|
|
"loss": 0.6848,
|
|
"step": 464
|
|
},
|
|
{
|
|
"epoch": 1.54,
|
|
"learning_rate": 4.021739130434782e-06,
|
|
"loss": 0.7221,
|
|
"step": 465
|
|
},
|
|
{
|
|
"epoch": 1.55,
|
|
"learning_rate": 4.016304347826087e-06,
|
|
"loss": 0.698,
|
|
"step": 466
|
|
},
|
|
{
|
|
"epoch": 1.55,
|
|
"learning_rate": 4.0108695652173915e-06,
|
|
"loss": 0.7091,
|
|
"step": 467
|
|
},
|
|
{
|
|
"epoch": 1.55,
|
|
"learning_rate": 4.005434782608696e-06,
|
|
"loss": 0.6842,
|
|
"step": 468
|
|
},
|
|
{
|
|
"epoch": 1.56,
|
|
"learning_rate": 4e-06,
|
|
"loss": 0.7131,
|
|
"step": 469
|
|
},
|
|
{
|
|
"epoch": 1.56,
|
|
"learning_rate": 3.994565217391305e-06,
|
|
"loss": 0.7156,
|
|
"step": 470
|
|
},
|
|
{
|
|
"epoch": 1.56,
|
|
"learning_rate": 3.989130434782608e-06,
|
|
"loss": 0.7034,
|
|
"step": 471
|
|
},
|
|
{
|
|
"epoch": 1.57,
|
|
"learning_rate": 3.983695652173913e-06,
|
|
"loss": 0.7122,
|
|
"step": 472
|
|
},
|
|
{
|
|
"epoch": 1.57,
|
|
"learning_rate": 3.978260869565217e-06,
|
|
"loss": 0.6918,
|
|
"step": 473
|
|
},
|
|
{
|
|
"epoch": 1.57,
|
|
"learning_rate": 3.9728260869565216e-06,
|
|
"loss": 0.6963,
|
|
"step": 474
|
|
},
|
|
{
|
|
"epoch": 1.58,
|
|
"learning_rate": 3.967391304347827e-06,
|
|
"loss": 0.7275,
|
|
"step": 475
|
|
},
|
|
{
|
|
"epoch": 1.58,
|
|
"learning_rate": 3.961956521739131e-06,
|
|
"loss": 0.6986,
|
|
"step": 476
|
|
},
|
|
{
|
|
"epoch": 1.58,
|
|
"learning_rate": 3.956521739130435e-06,
|
|
"loss": 0.7079,
|
|
"step": 477
|
|
},
|
|
{
|
|
"epoch": 1.59,
|
|
"learning_rate": 3.951086956521739e-06,
|
|
"loss": 0.7087,
|
|
"step": 478
|
|
},
|
|
{
|
|
"epoch": 1.59,
|
|
"learning_rate": 3.945652173913044e-06,
|
|
"loss": 0.7108,
|
|
"step": 479
|
|
},
|
|
{
|
|
"epoch": 1.59,
|
|
"learning_rate": 3.9402173913043475e-06,
|
|
"loss": 0.6959,
|
|
"step": 480
|
|
},
|
|
{
|
|
"epoch": 1.6,
|
|
"learning_rate": 3.9347826086956525e-06,
|
|
"loss": 0.6995,
|
|
"step": 481
|
|
},
|
|
{
|
|
"epoch": 1.6,
|
|
"learning_rate": 3.929347826086957e-06,
|
|
"loss": 0.708,
|
|
"step": 482
|
|
},
|
|
{
|
|
"epoch": 1.6,
|
|
"learning_rate": 3.923913043478261e-06,
|
|
"loss": 0.7006,
|
|
"step": 483
|
|
},
|
|
{
|
|
"epoch": 1.61,
|
|
"learning_rate": 3.918478260869565e-06,
|
|
"loss": 0.6958,
|
|
"step": 484
|
|
},
|
|
{
|
|
"epoch": 1.61,
|
|
"learning_rate": 3.91304347826087e-06,
|
|
"loss": 0.6967,
|
|
"step": 485
|
|
},
|
|
{
|
|
"epoch": 1.61,
|
|
"learning_rate": 3.907608695652173e-06,
|
|
"loss": 0.7078,
|
|
"step": 486
|
|
},
|
|
{
|
|
"epoch": 1.62,
|
|
"learning_rate": 3.9021739130434784e-06,
|
|
"loss": 0.7017,
|
|
"step": 487
|
|
},
|
|
{
|
|
"epoch": 1.62,
|
|
"learning_rate": 3.896739130434783e-06,
|
|
"loss": 0.6794,
|
|
"step": 488
|
|
},
|
|
{
|
|
"epoch": 1.62,
|
|
"eval_loss": 0.7612941861152649,
|
|
"eval_runtime": 17.6201,
|
|
"eval_samples_per_second": 129.511,
|
|
"eval_steps_per_second": 5.448,
|
|
"step": 488
|
|
},
|
|
{
|
|
"epoch": 1.62,
|
|
"learning_rate": 3.891304347826087e-06,
|
|
"loss": 0.6965,
|
|
"step": 489
|
|
},
|
|
{
|
|
"epoch": 1.63,
|
|
"learning_rate": 3.885869565217392e-06,
|
|
"loss": 0.676,
|
|
"step": 490
|
|
},
|
|
{
|
|
"epoch": 1.63,
|
|
"learning_rate": 3.880434782608696e-06,
|
|
"loss": 0.6866,
|
|
"step": 491
|
|
},
|
|
{
|
|
"epoch": 1.63,
|
|
"learning_rate": 3.875e-06,
|
|
"loss": 0.6896,
|
|
"step": 492
|
|
},
|
|
{
|
|
"epoch": 1.64,
|
|
"learning_rate": 3.869565217391304e-06,
|
|
"loss": 0.7011,
|
|
"step": 493
|
|
},
|
|
{
|
|
"epoch": 1.64,
|
|
"learning_rate": 3.864130434782609e-06,
|
|
"loss": 0.6953,
|
|
"step": 494
|
|
},
|
|
{
|
|
"epoch": 1.64,
|
|
"learning_rate": 3.858695652173913e-06,
|
|
"loss": 0.7148,
|
|
"step": 495
|
|
},
|
|
{
|
|
"epoch": 1.65,
|
|
"learning_rate": 3.853260869565218e-06,
|
|
"loss": 0.6997,
|
|
"step": 496
|
|
},
|
|
{
|
|
"epoch": 1.65,
|
|
"learning_rate": 3.847826086956522e-06,
|
|
"loss": 0.71,
|
|
"step": 497
|
|
},
|
|
{
|
|
"epoch": 1.65,
|
|
"learning_rate": 3.842391304347826e-06,
|
|
"loss": 0.6852,
|
|
"step": 498
|
|
},
|
|
{
|
|
"epoch": 1.66,
|
|
"learning_rate": 3.83695652173913e-06,
|
|
"loss": 0.6968,
|
|
"step": 499
|
|
},
|
|
{
|
|
"epoch": 1.66,
|
|
"learning_rate": 3.831521739130435e-06,
|
|
"loss": 0.6925,
|
|
"step": 500
|
|
},
|
|
{
|
|
"epoch": 1.66,
|
|
"learning_rate": 3.826086956521739e-06,
|
|
"loss": 0.717,
|
|
"step": 501
|
|
},
|
|
{
|
|
"epoch": 1.67,
|
|
"learning_rate": 3.820652173913044e-06,
|
|
"loss": 0.6857,
|
|
"step": 502
|
|
},
|
|
{
|
|
"epoch": 1.67,
|
|
"learning_rate": 3.815217391304348e-06,
|
|
"loss": 0.6861,
|
|
"step": 503
|
|
},
|
|
{
|
|
"epoch": 1.67,
|
|
"learning_rate": 3.809782608695652e-06,
|
|
"loss": 0.7046,
|
|
"step": 504
|
|
},
|
|
{
|
|
"epoch": 1.67,
|
|
"learning_rate": 3.804347826086957e-06,
|
|
"loss": 0.6928,
|
|
"step": 505
|
|
},
|
|
{
|
|
"epoch": 1.68,
|
|
"learning_rate": 3.798913043478261e-06,
|
|
"loss": 0.6995,
|
|
"step": 506
|
|
},
|
|
{
|
|
"epoch": 1.68,
|
|
"learning_rate": 3.7934782608695654e-06,
|
|
"loss": 0.7215,
|
|
"step": 507
|
|
},
|
|
{
|
|
"epoch": 1.68,
|
|
"learning_rate": 3.7880434782608696e-06,
|
|
"loss": 0.6872,
|
|
"step": 508
|
|
},
|
|
{
|
|
"epoch": 1.69,
|
|
"learning_rate": 3.782608695652174e-06,
|
|
"loss": 0.7018,
|
|
"step": 509
|
|
},
|
|
{
|
|
"epoch": 1.69,
|
|
"learning_rate": 3.7771739130434784e-06,
|
|
"loss": 0.702,
|
|
"step": 510
|
|
},
|
|
{
|
|
"epoch": 1.69,
|
|
"learning_rate": 3.771739130434783e-06,
|
|
"loss": 0.6838,
|
|
"step": 511
|
|
},
|
|
{
|
|
"epoch": 1.7,
|
|
"learning_rate": 3.7663043478260867e-06,
|
|
"loss": 0.7072,
|
|
"step": 512
|
|
},
|
|
{
|
|
"epoch": 1.7,
|
|
"learning_rate": 3.7608695652173913e-06,
|
|
"loss": 0.7073,
|
|
"step": 513
|
|
},
|
|
{
|
|
"epoch": 1.7,
|
|
"learning_rate": 3.7554347826086955e-06,
|
|
"loss": 0.7026,
|
|
"step": 514
|
|
},
|
|
{
|
|
"epoch": 1.71,
|
|
"learning_rate": 3.75e-06,
|
|
"loss": 0.7,
|
|
"step": 515
|
|
},
|
|
{
|
|
"epoch": 1.71,
|
|
"learning_rate": 3.7445652173913047e-06,
|
|
"loss": 0.7,
|
|
"step": 516
|
|
},
|
|
{
|
|
"epoch": 1.71,
|
|
"learning_rate": 3.739130434782609e-06,
|
|
"loss": 0.6895,
|
|
"step": 517
|
|
},
|
|
{
|
|
"epoch": 1.72,
|
|
"learning_rate": 3.7336956521739135e-06,
|
|
"loss": 0.704,
|
|
"step": 518
|
|
},
|
|
{
|
|
"epoch": 1.72,
|
|
"learning_rate": 3.7282608695652172e-06,
|
|
"loss": 0.7022,
|
|
"step": 519
|
|
},
|
|
{
|
|
"epoch": 1.72,
|
|
"learning_rate": 3.722826086956522e-06,
|
|
"loss": 0.7192,
|
|
"step": 520
|
|
},
|
|
{
|
|
"epoch": 1.73,
|
|
"learning_rate": 3.717391304347826e-06,
|
|
"loss": 0.6944,
|
|
"step": 521
|
|
},
|
|
{
|
|
"epoch": 1.73,
|
|
"learning_rate": 3.7119565217391306e-06,
|
|
"loss": 0.6886,
|
|
"step": 522
|
|
},
|
|
{
|
|
"epoch": 1.73,
|
|
"learning_rate": 3.706521739130435e-06,
|
|
"loss": 0.694,
|
|
"step": 523
|
|
},
|
|
{
|
|
"epoch": 1.74,
|
|
"learning_rate": 3.7010869565217394e-06,
|
|
"loss": 0.7102,
|
|
"step": 524
|
|
},
|
|
{
|
|
"epoch": 1.74,
|
|
"learning_rate": 3.695652173913043e-06,
|
|
"loss": 0.7203,
|
|
"step": 525
|
|
},
|
|
{
|
|
"epoch": 1.74,
|
|
"learning_rate": 3.690217391304348e-06,
|
|
"loss": 0.6784,
|
|
"step": 526
|
|
},
|
|
{
|
|
"epoch": 1.75,
|
|
"learning_rate": 3.684782608695652e-06,
|
|
"loss": 0.6962,
|
|
"step": 527
|
|
},
|
|
{
|
|
"epoch": 1.75,
|
|
"learning_rate": 3.6793478260869565e-06,
|
|
"loss": 0.6813,
|
|
"step": 528
|
|
},
|
|
{
|
|
"epoch": 1.75,
|
|
"learning_rate": 3.673913043478261e-06,
|
|
"loss": 0.6897,
|
|
"step": 529
|
|
},
|
|
{
|
|
"epoch": 1.76,
|
|
"learning_rate": 3.6684782608695653e-06,
|
|
"loss": 0.7197,
|
|
"step": 530
|
|
},
|
|
{
|
|
"epoch": 1.76,
|
|
"learning_rate": 3.66304347826087e-06,
|
|
"loss": 0.6985,
|
|
"step": 531
|
|
},
|
|
{
|
|
"epoch": 1.76,
|
|
"learning_rate": 3.657608695652174e-06,
|
|
"loss": 0.6807,
|
|
"step": 532
|
|
},
|
|
{
|
|
"epoch": 1.77,
|
|
"learning_rate": 3.6521739130434787e-06,
|
|
"loss": 0.7106,
|
|
"step": 533
|
|
},
|
|
{
|
|
"epoch": 1.77,
|
|
"learning_rate": 3.6467391304347825e-06,
|
|
"loss": 0.6843,
|
|
"step": 534
|
|
},
|
|
{
|
|
"epoch": 1.77,
|
|
"learning_rate": 3.641304347826087e-06,
|
|
"loss": 0.6941,
|
|
"step": 535
|
|
},
|
|
{
|
|
"epoch": 1.78,
|
|
"learning_rate": 3.6358695652173912e-06,
|
|
"loss": 0.6739,
|
|
"step": 536
|
|
},
|
|
{
|
|
"epoch": 1.78,
|
|
"learning_rate": 3.630434782608696e-06,
|
|
"loss": 0.7088,
|
|
"step": 537
|
|
},
|
|
{
|
|
"epoch": 1.78,
|
|
"learning_rate": 3.625e-06,
|
|
"loss": 0.7152,
|
|
"step": 538
|
|
},
|
|
{
|
|
"epoch": 1.79,
|
|
"learning_rate": 3.6195652173913046e-06,
|
|
"loss": 0.6972,
|
|
"step": 539
|
|
},
|
|
{
|
|
"epoch": 1.79,
|
|
"learning_rate": 3.6141304347826084e-06,
|
|
"loss": 0.6745,
|
|
"step": 540
|
|
},
|
|
{
|
|
"epoch": 1.79,
|
|
"learning_rate": 3.608695652173913e-06,
|
|
"loss": 0.7125,
|
|
"step": 541
|
|
},
|
|
{
|
|
"epoch": 1.8,
|
|
"learning_rate": 3.603260869565217e-06,
|
|
"loss": 0.7072,
|
|
"step": 542
|
|
},
|
|
{
|
|
"epoch": 1.8,
|
|
"learning_rate": 3.5978260869565218e-06,
|
|
"loss": 0.6903,
|
|
"step": 543
|
|
},
|
|
{
|
|
"epoch": 1.8,
|
|
"learning_rate": 3.5923913043478264e-06,
|
|
"loss": 0.6906,
|
|
"step": 544
|
|
},
|
|
{
|
|
"epoch": 1.81,
|
|
"learning_rate": 3.5869565217391305e-06,
|
|
"loss": 0.7101,
|
|
"step": 545
|
|
},
|
|
{
|
|
"epoch": 1.81,
|
|
"learning_rate": 3.581521739130435e-06,
|
|
"loss": 0.7097,
|
|
"step": 546
|
|
},
|
|
{
|
|
"epoch": 1.81,
|
|
"learning_rate": 3.5760869565217393e-06,
|
|
"loss": 0.7042,
|
|
"step": 547
|
|
},
|
|
{
|
|
"epoch": 1.82,
|
|
"learning_rate": 3.570652173913044e-06,
|
|
"loss": 0.7147,
|
|
"step": 548
|
|
},
|
|
{
|
|
"epoch": 1.82,
|
|
"learning_rate": 3.5652173913043477e-06,
|
|
"loss": 0.7039,
|
|
"step": 549
|
|
},
|
|
{
|
|
"epoch": 1.82,
|
|
"eval_loss": 0.7584934830665588,
|
|
"eval_runtime": 17.6166,
|
|
"eval_samples_per_second": 129.537,
|
|
"eval_steps_per_second": 5.449,
|
|
"step": 549
|
|
},
|
|
{
|
|
"epoch": 1.82,
|
|
"learning_rate": 3.5597826086956523e-06,
|
|
"loss": 0.6864,
|
|
"step": 550
|
|
},
|
|
{
|
|
"epoch": 1.83,
|
|
"learning_rate": 3.5543478260869565e-06,
|
|
"loss": 0.6858,
|
|
"step": 551
|
|
},
|
|
{
|
|
"epoch": 1.83,
|
|
"learning_rate": 3.548913043478261e-06,
|
|
"loss": 0.7076,
|
|
"step": 552
|
|
},
|
|
{
|
|
"epoch": 1.83,
|
|
"learning_rate": 3.5434782608695652e-06,
|
|
"loss": 0.6965,
|
|
"step": 553
|
|
},
|
|
{
|
|
"epoch": 1.84,
|
|
"learning_rate": 3.53804347826087e-06,
|
|
"loss": 0.6954,
|
|
"step": 554
|
|
},
|
|
{
|
|
"epoch": 1.84,
|
|
"learning_rate": 3.5326086956521736e-06,
|
|
"loss": 0.6934,
|
|
"step": 555
|
|
},
|
|
{
|
|
"epoch": 1.84,
|
|
"learning_rate": 3.527173913043478e-06,
|
|
"loss": 0.7159,
|
|
"step": 556
|
|
},
|
|
{
|
|
"epoch": 1.85,
|
|
"learning_rate": 3.521739130434783e-06,
|
|
"loss": 0.6946,
|
|
"step": 557
|
|
},
|
|
{
|
|
"epoch": 1.85,
|
|
"learning_rate": 3.516304347826087e-06,
|
|
"loss": 0.6866,
|
|
"step": 558
|
|
},
|
|
{
|
|
"epoch": 1.85,
|
|
"learning_rate": 3.5108695652173916e-06,
|
|
"loss": 0.704,
|
|
"step": 559
|
|
},
|
|
{
|
|
"epoch": 1.86,
|
|
"learning_rate": 3.5054347826086958e-06,
|
|
"loss": 0.6896,
|
|
"step": 560
|
|
},
|
|
{
|
|
"epoch": 1.86,
|
|
"learning_rate": 3.5000000000000004e-06,
|
|
"loss": 0.6947,
|
|
"step": 561
|
|
},
|
|
{
|
|
"epoch": 1.86,
|
|
"learning_rate": 3.494565217391304e-06,
|
|
"loss": 0.6993,
|
|
"step": 562
|
|
},
|
|
{
|
|
"epoch": 1.87,
|
|
"learning_rate": 3.489130434782609e-06,
|
|
"loss": 0.7046,
|
|
"step": 563
|
|
},
|
|
{
|
|
"epoch": 1.87,
|
|
"learning_rate": 3.483695652173913e-06,
|
|
"loss": 0.7037,
|
|
"step": 564
|
|
},
|
|
{
|
|
"epoch": 1.87,
|
|
"learning_rate": 3.4782608695652175e-06,
|
|
"loss": 0.7024,
|
|
"step": 565
|
|
},
|
|
{
|
|
"epoch": 1.88,
|
|
"learning_rate": 3.4728260869565217e-06,
|
|
"loss": 0.7186,
|
|
"step": 566
|
|
},
|
|
{
|
|
"epoch": 1.88,
|
|
"learning_rate": 3.4673913043478263e-06,
|
|
"loss": 0.6894,
|
|
"step": 567
|
|
},
|
|
{
|
|
"epoch": 1.88,
|
|
"learning_rate": 3.46195652173913e-06,
|
|
"loss": 0.7006,
|
|
"step": 568
|
|
},
|
|
{
|
|
"epoch": 1.89,
|
|
"learning_rate": 3.456521739130435e-06,
|
|
"loss": 0.6949,
|
|
"step": 569
|
|
},
|
|
{
|
|
"epoch": 1.89,
|
|
"learning_rate": 3.451086956521739e-06,
|
|
"loss": 0.6846,
|
|
"step": 570
|
|
},
|
|
{
|
|
"epoch": 1.89,
|
|
"learning_rate": 3.4456521739130434e-06,
|
|
"loss": 0.7118,
|
|
"step": 571
|
|
},
|
|
{
|
|
"epoch": 1.9,
|
|
"learning_rate": 3.440217391304348e-06,
|
|
"loss": 0.6945,
|
|
"step": 572
|
|
},
|
|
{
|
|
"epoch": 1.9,
|
|
"learning_rate": 3.4347826086956522e-06,
|
|
"loss": 0.6869,
|
|
"step": 573
|
|
},
|
|
{
|
|
"epoch": 1.9,
|
|
"learning_rate": 3.429347826086957e-06,
|
|
"loss": 0.7181,
|
|
"step": 574
|
|
},
|
|
{
|
|
"epoch": 1.91,
|
|
"learning_rate": 3.423913043478261e-06,
|
|
"loss": 0.7133,
|
|
"step": 575
|
|
},
|
|
{
|
|
"epoch": 1.91,
|
|
"learning_rate": 3.4184782608695656e-06,
|
|
"loss": 0.7068,
|
|
"step": 576
|
|
},
|
|
{
|
|
"epoch": 1.91,
|
|
"learning_rate": 3.4130434782608694e-06,
|
|
"loss": 0.6977,
|
|
"step": 577
|
|
},
|
|
{
|
|
"epoch": 1.92,
|
|
"learning_rate": 3.407608695652174e-06,
|
|
"loss": 0.6988,
|
|
"step": 578
|
|
},
|
|
{
|
|
"epoch": 1.92,
|
|
"learning_rate": 3.402173913043478e-06,
|
|
"loss": 0.7052,
|
|
"step": 579
|
|
},
|
|
{
|
|
"epoch": 1.92,
|
|
"learning_rate": 3.3967391304347827e-06,
|
|
"loss": 0.7249,
|
|
"step": 580
|
|
},
|
|
{
|
|
"epoch": 1.93,
|
|
"learning_rate": 3.391304347826087e-06,
|
|
"loss": 0.7198,
|
|
"step": 581
|
|
},
|
|
{
|
|
"epoch": 1.93,
|
|
"learning_rate": 3.3858695652173915e-06,
|
|
"loss": 0.6859,
|
|
"step": 582
|
|
},
|
|
{
|
|
"epoch": 1.93,
|
|
"learning_rate": 3.3804347826086953e-06,
|
|
"loss": 0.683,
|
|
"step": 583
|
|
},
|
|
{
|
|
"epoch": 1.94,
|
|
"learning_rate": 3.375e-06,
|
|
"loss": 0.6997,
|
|
"step": 584
|
|
},
|
|
{
|
|
"epoch": 1.94,
|
|
"learning_rate": 3.369565217391305e-06,
|
|
"loss": 0.6758,
|
|
"step": 585
|
|
},
|
|
{
|
|
"epoch": 1.94,
|
|
"learning_rate": 3.3641304347826087e-06,
|
|
"loss": 0.7062,
|
|
"step": 586
|
|
},
|
|
{
|
|
"epoch": 1.95,
|
|
"learning_rate": 3.3586956521739133e-06,
|
|
"loss": 0.6972,
|
|
"step": 587
|
|
},
|
|
{
|
|
"epoch": 1.95,
|
|
"learning_rate": 3.3532608695652174e-06,
|
|
"loss": 0.7078,
|
|
"step": 588
|
|
},
|
|
{
|
|
"epoch": 1.95,
|
|
"learning_rate": 3.347826086956522e-06,
|
|
"loss": 0.7176,
|
|
"step": 589
|
|
},
|
|
{
|
|
"epoch": 1.96,
|
|
"learning_rate": 3.3423913043478262e-06,
|
|
"loss": 0.6931,
|
|
"step": 590
|
|
},
|
|
{
|
|
"epoch": 1.96,
|
|
"learning_rate": 3.336956521739131e-06,
|
|
"loss": 0.6981,
|
|
"step": 591
|
|
},
|
|
{
|
|
"epoch": 1.96,
|
|
"learning_rate": 3.3315217391304346e-06,
|
|
"loss": 0.7222,
|
|
"step": 592
|
|
},
|
|
{
|
|
"epoch": 1.97,
|
|
"learning_rate": 3.326086956521739e-06,
|
|
"loss": 0.6859,
|
|
"step": 593
|
|
},
|
|
{
|
|
"epoch": 1.97,
|
|
"learning_rate": 3.3206521739130434e-06,
|
|
"loss": 0.6878,
|
|
"step": 594
|
|
},
|
|
{
|
|
"epoch": 1.97,
|
|
"learning_rate": 3.315217391304348e-06,
|
|
"loss": 0.6986,
|
|
"step": 595
|
|
},
|
|
{
|
|
"epoch": 1.98,
|
|
"learning_rate": 3.309782608695652e-06,
|
|
"loss": 0.6882,
|
|
"step": 596
|
|
},
|
|
{
|
|
"epoch": 1.98,
|
|
"learning_rate": 3.3043478260869567e-06,
|
|
"loss": 0.7049,
|
|
"step": 597
|
|
},
|
|
{
|
|
"epoch": 1.98,
|
|
"learning_rate": 3.2989130434782613e-06,
|
|
"loss": 0.6961,
|
|
"step": 598
|
|
},
|
|
{
|
|
"epoch": 1.99,
|
|
"learning_rate": 3.293478260869565e-06,
|
|
"loss": 0.6907,
|
|
"step": 599
|
|
},
|
|
{
|
|
"epoch": 1.99,
|
|
"learning_rate": 3.28804347826087e-06,
|
|
"loss": 0.7177,
|
|
"step": 600
|
|
},
|
|
{
|
|
"epoch": 1.99,
|
|
"learning_rate": 3.282608695652174e-06,
|
|
"loss": 0.6917,
|
|
"step": 601
|
|
},
|
|
{
|
|
"epoch": 2.0,
|
|
"learning_rate": 3.2771739130434785e-06,
|
|
"loss": 0.6966,
|
|
"step": 602
|
|
},
|
|
{
|
|
"epoch": 2.0,
|
|
"learning_rate": 3.2717391304347827e-06,
|
|
"loss": 0.7036,
|
|
"step": 603
|
|
},
|
|
{
|
|
"epoch": 2.0,
|
|
"learning_rate": 3.2663043478260873e-06,
|
|
"loss": 0.6433,
|
|
"step": 604
|
|
},
|
|
{
|
|
"epoch": 2.01,
|
|
"learning_rate": 3.260869565217391e-06,
|
|
"loss": 0.6296,
|
|
"step": 605
|
|
},
|
|
{
|
|
"epoch": 2.01,
|
|
"learning_rate": 3.255434782608696e-06,
|
|
"loss": 0.6323,
|
|
"step": 606
|
|
},
|
|
{
|
|
"epoch": 2.01,
|
|
"learning_rate": 3.25e-06,
|
|
"loss": 0.64,
|
|
"step": 607
|
|
},
|
|
{
|
|
"epoch": 2.02,
|
|
"learning_rate": 3.2445652173913044e-06,
|
|
"loss": 0.6384,
|
|
"step": 608
|
|
},
|
|
{
|
|
"epoch": 2.02,
|
|
"learning_rate": 3.2391304347826086e-06,
|
|
"loss": 0.6355,
|
|
"step": 609
|
|
},
|
|
{
|
|
"epoch": 2.02,
|
|
"learning_rate": 3.233695652173913e-06,
|
|
"loss": 0.6299,
|
|
"step": 610
|
|
},
|
|
{
|
|
"epoch": 2.02,
|
|
"eval_loss": 0.7699483036994934,
|
|
"eval_runtime": 17.6263,
|
|
"eval_samples_per_second": 129.466,
|
|
"eval_steps_per_second": 5.446,
|
|
"step": 610
|
|
},
|
|
{
|
|
"epoch": 2.03,
|
|
"learning_rate": 3.2282608695652174e-06,
|
|
"loss": 0.6174,
|
|
"step": 611
|
|
},
|
|
{
|
|
"epoch": 2.03,
|
|
"learning_rate": 3.222826086956522e-06,
|
|
"loss": 0.6296,
|
|
"step": 612
|
|
},
|
|
{
|
|
"epoch": 2.03,
|
|
"learning_rate": 3.2173913043478266e-06,
|
|
"loss": 0.6246,
|
|
"step": 613
|
|
},
|
|
{
|
|
"epoch": 2.04,
|
|
"learning_rate": 3.2119565217391303e-06,
|
|
"loss": 0.6361,
|
|
"step": 614
|
|
},
|
|
{
|
|
"epoch": 2.04,
|
|
"learning_rate": 3.206521739130435e-06,
|
|
"loss": 0.6313,
|
|
"step": 615
|
|
},
|
|
{
|
|
"epoch": 2.04,
|
|
"learning_rate": 3.201086956521739e-06,
|
|
"loss": 0.638,
|
|
"step": 616
|
|
},
|
|
{
|
|
"epoch": 2.05,
|
|
"learning_rate": 3.1956521739130437e-06,
|
|
"loss": 0.6223,
|
|
"step": 617
|
|
},
|
|
{
|
|
"epoch": 2.05,
|
|
"learning_rate": 3.190217391304348e-06,
|
|
"loss": 0.6325,
|
|
"step": 618
|
|
},
|
|
{
|
|
"epoch": 2.05,
|
|
"learning_rate": 3.1847826086956525e-06,
|
|
"loss": 0.6369,
|
|
"step": 619
|
|
},
|
|
{
|
|
"epoch": 2.06,
|
|
"learning_rate": 3.1793478260869562e-06,
|
|
"loss": 0.6295,
|
|
"step": 620
|
|
},
|
|
{
|
|
"epoch": 2.06,
|
|
"learning_rate": 3.173913043478261e-06,
|
|
"loss": 0.6335,
|
|
"step": 621
|
|
},
|
|
{
|
|
"epoch": 2.06,
|
|
"learning_rate": 3.168478260869565e-06,
|
|
"loss": 0.6275,
|
|
"step": 622
|
|
},
|
|
{
|
|
"epoch": 2.07,
|
|
"learning_rate": 3.1630434782608696e-06,
|
|
"loss": 0.6226,
|
|
"step": 623
|
|
},
|
|
{
|
|
"epoch": 2.07,
|
|
"learning_rate": 3.157608695652174e-06,
|
|
"loss": 0.6326,
|
|
"step": 624
|
|
},
|
|
{
|
|
"epoch": 2.07,
|
|
"learning_rate": 3.1521739130434784e-06,
|
|
"loss": 0.6292,
|
|
"step": 625
|
|
},
|
|
{
|
|
"epoch": 2.08,
|
|
"learning_rate": 3.146739130434783e-06,
|
|
"loss": 0.6375,
|
|
"step": 626
|
|
},
|
|
{
|
|
"epoch": 2.08,
|
|
"learning_rate": 3.141304347826087e-06,
|
|
"loss": 0.6404,
|
|
"step": 627
|
|
},
|
|
{
|
|
"epoch": 2.08,
|
|
"learning_rate": 3.135869565217392e-06,
|
|
"loss": 0.6348,
|
|
"step": 628
|
|
},
|
|
{
|
|
"epoch": 2.09,
|
|
"learning_rate": 3.1304347826086955e-06,
|
|
"loss": 0.6341,
|
|
"step": 629
|
|
},
|
|
{
|
|
"epoch": 2.09,
|
|
"learning_rate": 3.125e-06,
|
|
"loss": 0.6085,
|
|
"step": 630
|
|
},
|
|
{
|
|
"epoch": 2.09,
|
|
"learning_rate": 3.1195652173913043e-06,
|
|
"loss": 0.6096,
|
|
"step": 631
|
|
},
|
|
{
|
|
"epoch": 2.1,
|
|
"learning_rate": 3.114130434782609e-06,
|
|
"loss": 0.6322,
|
|
"step": 632
|
|
},
|
|
{
|
|
"epoch": 2.1,
|
|
"learning_rate": 3.108695652173913e-06,
|
|
"loss": 0.6418,
|
|
"step": 633
|
|
},
|
|
{
|
|
"epoch": 2.1,
|
|
"learning_rate": 3.1032608695652177e-06,
|
|
"loss": 0.6248,
|
|
"step": 634
|
|
},
|
|
{
|
|
"epoch": 2.11,
|
|
"learning_rate": 3.0978260869565215e-06,
|
|
"loss": 0.631,
|
|
"step": 635
|
|
},
|
|
{
|
|
"epoch": 2.11,
|
|
"learning_rate": 3.092391304347826e-06,
|
|
"loss": 0.6203,
|
|
"step": 636
|
|
},
|
|
{
|
|
"epoch": 2.11,
|
|
"learning_rate": 3.0869565217391302e-06,
|
|
"loss": 0.6167,
|
|
"step": 637
|
|
},
|
|
{
|
|
"epoch": 2.12,
|
|
"learning_rate": 3.081521739130435e-06,
|
|
"loss": 0.6372,
|
|
"step": 638
|
|
},
|
|
{
|
|
"epoch": 2.12,
|
|
"learning_rate": 3.076086956521739e-06,
|
|
"loss": 0.6196,
|
|
"step": 639
|
|
},
|
|
{
|
|
"epoch": 2.12,
|
|
"learning_rate": 3.0706521739130436e-06,
|
|
"loss": 0.6187,
|
|
"step": 640
|
|
},
|
|
{
|
|
"epoch": 2.13,
|
|
"learning_rate": 3.0652173913043482e-06,
|
|
"loss": 0.6264,
|
|
"step": 641
|
|
},
|
|
{
|
|
"epoch": 2.13,
|
|
"learning_rate": 3.059782608695652e-06,
|
|
"loss": 0.6285,
|
|
"step": 642
|
|
},
|
|
{
|
|
"epoch": 2.13,
|
|
"learning_rate": 3.054347826086957e-06,
|
|
"loss": 0.6237,
|
|
"step": 643
|
|
},
|
|
{
|
|
"epoch": 2.14,
|
|
"learning_rate": 3.0489130434782608e-06,
|
|
"loss": 0.6242,
|
|
"step": 644
|
|
},
|
|
{
|
|
"epoch": 2.14,
|
|
"learning_rate": 3.0434782608695654e-06,
|
|
"loss": 0.6337,
|
|
"step": 645
|
|
},
|
|
{
|
|
"epoch": 2.14,
|
|
"learning_rate": 3.0380434782608696e-06,
|
|
"loss": 0.635,
|
|
"step": 646
|
|
},
|
|
{
|
|
"epoch": 2.15,
|
|
"learning_rate": 3.032608695652174e-06,
|
|
"loss": 0.6448,
|
|
"step": 647
|
|
},
|
|
{
|
|
"epoch": 2.15,
|
|
"learning_rate": 3.0271739130434783e-06,
|
|
"loss": 0.6375,
|
|
"step": 648
|
|
},
|
|
{
|
|
"epoch": 2.15,
|
|
"learning_rate": 3.021739130434783e-06,
|
|
"loss": 0.6213,
|
|
"step": 649
|
|
},
|
|
{
|
|
"epoch": 2.16,
|
|
"learning_rate": 3.0163043478260867e-06,
|
|
"loss": 0.632,
|
|
"step": 650
|
|
},
|
|
{
|
|
"epoch": 2.16,
|
|
"learning_rate": 3.0108695652173913e-06,
|
|
"loss": 0.6336,
|
|
"step": 651
|
|
},
|
|
{
|
|
"epoch": 2.16,
|
|
"learning_rate": 3.0054347826086955e-06,
|
|
"loss": 0.6361,
|
|
"step": 652
|
|
},
|
|
{
|
|
"epoch": 2.17,
|
|
"learning_rate": 3e-06,
|
|
"loss": 0.6177,
|
|
"step": 653
|
|
},
|
|
{
|
|
"epoch": 2.17,
|
|
"learning_rate": 2.9945652173913043e-06,
|
|
"loss": 0.6497,
|
|
"step": 654
|
|
},
|
|
{
|
|
"epoch": 2.17,
|
|
"learning_rate": 2.989130434782609e-06,
|
|
"loss": 0.621,
|
|
"step": 655
|
|
},
|
|
{
|
|
"epoch": 2.18,
|
|
"learning_rate": 2.983695652173913e-06,
|
|
"loss": 0.6304,
|
|
"step": 656
|
|
},
|
|
{
|
|
"epoch": 2.18,
|
|
"learning_rate": 2.9782608695652172e-06,
|
|
"loss": 0.6386,
|
|
"step": 657
|
|
},
|
|
{
|
|
"epoch": 2.18,
|
|
"learning_rate": 2.972826086956522e-06,
|
|
"loss": 0.6339,
|
|
"step": 658
|
|
},
|
|
{
|
|
"epoch": 2.19,
|
|
"learning_rate": 2.967391304347826e-06,
|
|
"loss": 0.6371,
|
|
"step": 659
|
|
},
|
|
{
|
|
"epoch": 2.19,
|
|
"learning_rate": 2.9619565217391306e-06,
|
|
"loss": 0.6375,
|
|
"step": 660
|
|
},
|
|
{
|
|
"epoch": 2.19,
|
|
"learning_rate": 2.956521739130435e-06,
|
|
"loss": 0.6463,
|
|
"step": 661
|
|
},
|
|
{
|
|
"epoch": 2.2,
|
|
"learning_rate": 2.9510869565217394e-06,
|
|
"loss": 0.6149,
|
|
"step": 662
|
|
},
|
|
{
|
|
"epoch": 2.2,
|
|
"learning_rate": 2.9456521739130436e-06,
|
|
"loss": 0.6465,
|
|
"step": 663
|
|
},
|
|
{
|
|
"epoch": 2.2,
|
|
"learning_rate": 2.940217391304348e-06,
|
|
"loss": 0.6415,
|
|
"step": 664
|
|
},
|
|
{
|
|
"epoch": 2.21,
|
|
"learning_rate": 2.9347826086956523e-06,
|
|
"loss": 0.6055,
|
|
"step": 665
|
|
},
|
|
{
|
|
"epoch": 2.21,
|
|
"learning_rate": 2.9293478260869565e-06,
|
|
"loss": 0.6254,
|
|
"step": 666
|
|
},
|
|
{
|
|
"epoch": 2.21,
|
|
"learning_rate": 2.923913043478261e-06,
|
|
"loss": 0.6116,
|
|
"step": 667
|
|
},
|
|
{
|
|
"epoch": 2.22,
|
|
"learning_rate": 2.9184782608695653e-06,
|
|
"loss": 0.6262,
|
|
"step": 668
|
|
},
|
|
{
|
|
"epoch": 2.22,
|
|
"learning_rate": 2.9130434782608695e-06,
|
|
"loss": 0.6373,
|
|
"step": 669
|
|
},
|
|
{
|
|
"epoch": 2.22,
|
|
"learning_rate": 2.907608695652174e-06,
|
|
"loss": 0.6348,
|
|
"step": 670
|
|
},
|
|
{
|
|
"epoch": 2.23,
|
|
"learning_rate": 2.9021739130434783e-06,
|
|
"loss": 0.6199,
|
|
"step": 671
|
|
},
|
|
{
|
|
"epoch": 2.23,
|
|
"eval_loss": 0.7679464221000671,
|
|
"eval_runtime": 17.626,
|
|
"eval_samples_per_second": 129.468,
|
|
"eval_steps_per_second": 5.446,
|
|
"step": 671
|
|
},
|
|
{
|
|
"epoch": 2.23,
|
|
"learning_rate": 2.8967391304347824e-06,
|
|
"loss": 0.6139,
|
|
"step": 672
|
|
},
|
|
{
|
|
"epoch": 2.23,
|
|
"learning_rate": 2.891304347826087e-06,
|
|
"loss": 0.6286,
|
|
"step": 673
|
|
},
|
|
{
|
|
"epoch": 2.24,
|
|
"learning_rate": 2.8858695652173916e-06,
|
|
"loss": 0.652,
|
|
"step": 674
|
|
},
|
|
{
|
|
"epoch": 2.24,
|
|
"learning_rate": 2.880434782608696e-06,
|
|
"loss": 0.6246,
|
|
"step": 675
|
|
},
|
|
{
|
|
"epoch": 2.24,
|
|
"learning_rate": 2.875e-06,
|
|
"loss": 0.6274,
|
|
"step": 676
|
|
},
|
|
{
|
|
"epoch": 2.25,
|
|
"learning_rate": 2.8695652173913046e-06,
|
|
"loss": 0.6589,
|
|
"step": 677
|
|
},
|
|
{
|
|
"epoch": 2.25,
|
|
"learning_rate": 2.8641304347826088e-06,
|
|
"loss": 0.6505,
|
|
"step": 678
|
|
},
|
|
{
|
|
"epoch": 2.25,
|
|
"learning_rate": 2.858695652173913e-06,
|
|
"loss": 0.6239,
|
|
"step": 679
|
|
},
|
|
{
|
|
"epoch": 2.26,
|
|
"learning_rate": 2.8532608695652176e-06,
|
|
"loss": 0.6199,
|
|
"step": 680
|
|
},
|
|
{
|
|
"epoch": 2.26,
|
|
"learning_rate": 2.8478260869565217e-06,
|
|
"loss": 0.6372,
|
|
"step": 681
|
|
},
|
|
{
|
|
"epoch": 2.26,
|
|
"learning_rate": 2.842391304347826e-06,
|
|
"loss": 0.6313,
|
|
"step": 682
|
|
},
|
|
{
|
|
"epoch": 2.27,
|
|
"learning_rate": 2.8369565217391305e-06,
|
|
"loss": 0.6295,
|
|
"step": 683
|
|
},
|
|
{
|
|
"epoch": 2.27,
|
|
"learning_rate": 2.8315217391304347e-06,
|
|
"loss": 0.6319,
|
|
"step": 684
|
|
},
|
|
{
|
|
"epoch": 2.27,
|
|
"learning_rate": 2.8260869565217393e-06,
|
|
"loss": 0.6432,
|
|
"step": 685
|
|
},
|
|
{
|
|
"epoch": 2.28,
|
|
"learning_rate": 2.8206521739130435e-06,
|
|
"loss": 0.6367,
|
|
"step": 686
|
|
},
|
|
{
|
|
"epoch": 2.28,
|
|
"learning_rate": 2.8152173913043477e-06,
|
|
"loss": 0.6343,
|
|
"step": 687
|
|
},
|
|
{
|
|
"epoch": 2.28,
|
|
"learning_rate": 2.8097826086956523e-06,
|
|
"loss": 0.6238,
|
|
"step": 688
|
|
},
|
|
{
|
|
"epoch": 2.29,
|
|
"learning_rate": 2.804347826086957e-06,
|
|
"loss": 0.6409,
|
|
"step": 689
|
|
},
|
|
{
|
|
"epoch": 2.29,
|
|
"learning_rate": 2.798913043478261e-06,
|
|
"loss": 0.6267,
|
|
"step": 690
|
|
},
|
|
{
|
|
"epoch": 2.29,
|
|
"learning_rate": 2.7934782608695652e-06,
|
|
"loss": 0.6319,
|
|
"step": 691
|
|
},
|
|
{
|
|
"epoch": 2.3,
|
|
"learning_rate": 2.78804347826087e-06,
|
|
"loss": 0.6363,
|
|
"step": 692
|
|
},
|
|
{
|
|
"epoch": 2.3,
|
|
"learning_rate": 2.782608695652174e-06,
|
|
"loss": 0.6341,
|
|
"step": 693
|
|
},
|
|
{
|
|
"epoch": 2.3,
|
|
"learning_rate": 2.777173913043478e-06,
|
|
"loss": 0.6347,
|
|
"step": 694
|
|
},
|
|
{
|
|
"epoch": 2.31,
|
|
"learning_rate": 2.771739130434783e-06,
|
|
"loss": 0.6392,
|
|
"step": 695
|
|
},
|
|
{
|
|
"epoch": 2.31,
|
|
"learning_rate": 2.766304347826087e-06,
|
|
"loss": 0.6529,
|
|
"step": 696
|
|
},
|
|
{
|
|
"epoch": 2.31,
|
|
"learning_rate": 2.760869565217391e-06,
|
|
"loss": 0.6399,
|
|
"step": 697
|
|
},
|
|
{
|
|
"epoch": 2.32,
|
|
"learning_rate": 2.7554347826086957e-06,
|
|
"loss": 0.6168,
|
|
"step": 698
|
|
},
|
|
{
|
|
"epoch": 2.32,
|
|
"learning_rate": 2.75e-06,
|
|
"loss": 0.6356,
|
|
"step": 699
|
|
},
|
|
{
|
|
"epoch": 2.32,
|
|
"learning_rate": 2.744565217391304e-06,
|
|
"loss": 0.6543,
|
|
"step": 700
|
|
},
|
|
{
|
|
"epoch": 2.33,
|
|
"learning_rate": 2.7391304347826087e-06,
|
|
"loss": 0.6373,
|
|
"step": 701
|
|
},
|
|
{
|
|
"epoch": 2.33,
|
|
"learning_rate": 2.7336956521739133e-06,
|
|
"loss": 0.6348,
|
|
"step": 702
|
|
},
|
|
{
|
|
"epoch": 2.33,
|
|
"learning_rate": 2.7282608695652175e-06,
|
|
"loss": 0.6327,
|
|
"step": 703
|
|
},
|
|
{
|
|
"epoch": 2.33,
|
|
"learning_rate": 2.722826086956522e-06,
|
|
"loss": 0.6171,
|
|
"step": 704
|
|
},
|
|
{
|
|
"epoch": 2.34,
|
|
"learning_rate": 2.7173913043478263e-06,
|
|
"loss": 0.6216,
|
|
"step": 705
|
|
},
|
|
{
|
|
"epoch": 2.34,
|
|
"learning_rate": 2.7119565217391305e-06,
|
|
"loss": 0.6314,
|
|
"step": 706
|
|
},
|
|
{
|
|
"epoch": 2.34,
|
|
"learning_rate": 2.706521739130435e-06,
|
|
"loss": 0.6293,
|
|
"step": 707
|
|
},
|
|
{
|
|
"epoch": 2.35,
|
|
"learning_rate": 2.7010869565217392e-06,
|
|
"loss": 0.6322,
|
|
"step": 708
|
|
},
|
|
{
|
|
"epoch": 2.35,
|
|
"learning_rate": 2.6956521739130434e-06,
|
|
"loss": 0.6474,
|
|
"step": 709
|
|
},
|
|
{
|
|
"epoch": 2.35,
|
|
"learning_rate": 2.690217391304348e-06,
|
|
"loss": 0.6381,
|
|
"step": 710
|
|
},
|
|
{
|
|
"epoch": 2.36,
|
|
"learning_rate": 2.684782608695652e-06,
|
|
"loss": 0.6139,
|
|
"step": 711
|
|
},
|
|
{
|
|
"epoch": 2.36,
|
|
"learning_rate": 2.6793478260869564e-06,
|
|
"loss": 0.6432,
|
|
"step": 712
|
|
},
|
|
{
|
|
"epoch": 2.36,
|
|
"learning_rate": 2.673913043478261e-06,
|
|
"loss": 0.6316,
|
|
"step": 713
|
|
},
|
|
{
|
|
"epoch": 2.37,
|
|
"learning_rate": 2.668478260869565e-06,
|
|
"loss": 0.6414,
|
|
"step": 714
|
|
},
|
|
{
|
|
"epoch": 2.37,
|
|
"learning_rate": 2.6630434782608693e-06,
|
|
"loss": 0.637,
|
|
"step": 715
|
|
},
|
|
{
|
|
"epoch": 2.37,
|
|
"learning_rate": 2.657608695652174e-06,
|
|
"loss": 0.6375,
|
|
"step": 716
|
|
},
|
|
{
|
|
"epoch": 2.38,
|
|
"learning_rate": 2.6521739130434785e-06,
|
|
"loss": 0.6324,
|
|
"step": 717
|
|
},
|
|
{
|
|
"epoch": 2.38,
|
|
"learning_rate": 2.6467391304347827e-06,
|
|
"loss": 0.6417,
|
|
"step": 718
|
|
},
|
|
{
|
|
"epoch": 2.38,
|
|
"learning_rate": 2.641304347826087e-06,
|
|
"loss": 0.6176,
|
|
"step": 719
|
|
},
|
|
{
|
|
"epoch": 2.39,
|
|
"learning_rate": 2.6358695652173915e-06,
|
|
"loss": 0.6438,
|
|
"step": 720
|
|
},
|
|
{
|
|
"epoch": 2.39,
|
|
"learning_rate": 2.6304347826086957e-06,
|
|
"loss": 0.6375,
|
|
"step": 721
|
|
},
|
|
{
|
|
"epoch": 2.39,
|
|
"learning_rate": 2.6250000000000003e-06,
|
|
"loss": 0.6248,
|
|
"step": 722
|
|
},
|
|
{
|
|
"epoch": 2.4,
|
|
"learning_rate": 2.6195652173913045e-06,
|
|
"loss": 0.6337,
|
|
"step": 723
|
|
},
|
|
{
|
|
"epoch": 2.4,
|
|
"learning_rate": 2.6141304347826086e-06,
|
|
"loss": 0.6517,
|
|
"step": 724
|
|
},
|
|
{
|
|
"epoch": 2.4,
|
|
"learning_rate": 2.6086956521739132e-06,
|
|
"loss": 0.6313,
|
|
"step": 725
|
|
},
|
|
{
|
|
"epoch": 2.41,
|
|
"learning_rate": 2.6032608695652174e-06,
|
|
"loss": 0.6425,
|
|
"step": 726
|
|
},
|
|
{
|
|
"epoch": 2.41,
|
|
"learning_rate": 2.5978260869565216e-06,
|
|
"loss": 0.6203,
|
|
"step": 727
|
|
},
|
|
{
|
|
"epoch": 2.41,
|
|
"learning_rate": 2.592391304347826e-06,
|
|
"loss": 0.6208,
|
|
"step": 728
|
|
},
|
|
{
|
|
"epoch": 2.42,
|
|
"learning_rate": 2.586956521739131e-06,
|
|
"loss": 0.6332,
|
|
"step": 729
|
|
},
|
|
{
|
|
"epoch": 2.42,
|
|
"learning_rate": 2.581521739130435e-06,
|
|
"loss": 0.6229,
|
|
"step": 730
|
|
},
|
|
{
|
|
"epoch": 2.42,
|
|
"learning_rate": 2.576086956521739e-06,
|
|
"loss": 0.6412,
|
|
"step": 731
|
|
},
|
|
{
|
|
"epoch": 2.43,
|
|
"learning_rate": 2.5706521739130438e-06,
|
|
"loss": 0.6498,
|
|
"step": 732
|
|
},
|
|
{
|
|
"epoch": 2.43,
|
|
"eval_loss": 0.768299400806427,
|
|
"eval_runtime": 17.6195,
|
|
"eval_samples_per_second": 129.516,
|
|
"eval_steps_per_second": 5.449,
|
|
"step": 732
|
|
},
|
|
{
|
|
"epoch": 2.43,
|
|
"learning_rate": 2.565217391304348e-06,
|
|
"loss": 0.6453,
|
|
"step": 733
|
|
},
|
|
{
|
|
"epoch": 2.43,
|
|
"learning_rate": 2.559782608695652e-06,
|
|
"loss": 0.6359,
|
|
"step": 734
|
|
},
|
|
{
|
|
"epoch": 2.44,
|
|
"learning_rate": 2.5543478260869567e-06,
|
|
"loss": 0.6303,
|
|
"step": 735
|
|
},
|
|
{
|
|
"epoch": 2.44,
|
|
"learning_rate": 2.548913043478261e-06,
|
|
"loss": 0.6458,
|
|
"step": 736
|
|
},
|
|
{
|
|
"epoch": 2.44,
|
|
"learning_rate": 2.543478260869565e-06,
|
|
"loss": 0.6565,
|
|
"step": 737
|
|
},
|
|
{
|
|
"epoch": 2.45,
|
|
"learning_rate": 2.5380434782608697e-06,
|
|
"loss": 0.6202,
|
|
"step": 738
|
|
},
|
|
{
|
|
"epoch": 2.45,
|
|
"learning_rate": 2.532608695652174e-06,
|
|
"loss": 0.6392,
|
|
"step": 739
|
|
},
|
|
{
|
|
"epoch": 2.45,
|
|
"learning_rate": 2.527173913043478e-06,
|
|
"loss": 0.6466,
|
|
"step": 740
|
|
},
|
|
{
|
|
"epoch": 2.46,
|
|
"learning_rate": 2.5217391304347826e-06,
|
|
"loss": 0.62,
|
|
"step": 741
|
|
},
|
|
{
|
|
"epoch": 2.46,
|
|
"learning_rate": 2.516304347826087e-06,
|
|
"loss": 0.6353,
|
|
"step": 742
|
|
},
|
|
{
|
|
"epoch": 2.46,
|
|
"learning_rate": 2.5108695652173914e-06,
|
|
"loss": 0.637,
|
|
"step": 743
|
|
},
|
|
{
|
|
"epoch": 2.47,
|
|
"learning_rate": 2.505434782608696e-06,
|
|
"loss": 0.63,
|
|
"step": 744
|
|
},
|
|
{
|
|
"epoch": 2.47,
|
|
"learning_rate": 2.5e-06,
|
|
"loss": 0.6419,
|
|
"step": 745
|
|
},
|
|
{
|
|
"epoch": 2.47,
|
|
"learning_rate": 2.4945652173913044e-06,
|
|
"loss": 0.6419,
|
|
"step": 746
|
|
},
|
|
{
|
|
"epoch": 2.48,
|
|
"learning_rate": 2.489130434782609e-06,
|
|
"loss": 0.6446,
|
|
"step": 747
|
|
},
|
|
{
|
|
"epoch": 2.48,
|
|
"learning_rate": 2.483695652173913e-06,
|
|
"loss": 0.6446,
|
|
"step": 748
|
|
},
|
|
{
|
|
"epoch": 2.48,
|
|
"learning_rate": 2.4782608695652173e-06,
|
|
"loss": 0.6272,
|
|
"step": 749
|
|
},
|
|
{
|
|
"epoch": 2.49,
|
|
"learning_rate": 2.472826086956522e-06,
|
|
"loss": 0.6417,
|
|
"step": 750
|
|
},
|
|
{
|
|
"epoch": 2.49,
|
|
"learning_rate": 2.467391304347826e-06,
|
|
"loss": 0.6328,
|
|
"step": 751
|
|
},
|
|
{
|
|
"epoch": 2.49,
|
|
"learning_rate": 2.4619565217391303e-06,
|
|
"loss": 0.6426,
|
|
"step": 752
|
|
},
|
|
{
|
|
"epoch": 2.5,
|
|
"learning_rate": 2.456521739130435e-06,
|
|
"loss": 0.6489,
|
|
"step": 753
|
|
},
|
|
{
|
|
"epoch": 2.5,
|
|
"learning_rate": 2.451086956521739e-06,
|
|
"loss": 0.6299,
|
|
"step": 754
|
|
},
|
|
{
|
|
"epoch": 2.5,
|
|
"learning_rate": 2.4456521739130433e-06,
|
|
"loss": 0.6457,
|
|
"step": 755
|
|
},
|
|
{
|
|
"epoch": 2.51,
|
|
"learning_rate": 2.440217391304348e-06,
|
|
"loss": 0.6306,
|
|
"step": 756
|
|
},
|
|
{
|
|
"epoch": 2.51,
|
|
"learning_rate": 2.4347826086956525e-06,
|
|
"loss": 0.639,
|
|
"step": 757
|
|
},
|
|
{
|
|
"epoch": 2.51,
|
|
"learning_rate": 2.4293478260869566e-06,
|
|
"loss": 0.6408,
|
|
"step": 758
|
|
},
|
|
{
|
|
"epoch": 2.52,
|
|
"learning_rate": 2.423913043478261e-06,
|
|
"loss": 0.64,
|
|
"step": 759
|
|
},
|
|
{
|
|
"epoch": 2.52,
|
|
"learning_rate": 2.4184782608695654e-06,
|
|
"loss": 0.6249,
|
|
"step": 760
|
|
},
|
|
{
|
|
"epoch": 2.52,
|
|
"learning_rate": 2.4130434782608696e-06,
|
|
"loss": 0.6382,
|
|
"step": 761
|
|
},
|
|
{
|
|
"epoch": 2.53,
|
|
"learning_rate": 2.407608695652174e-06,
|
|
"loss": 0.6328,
|
|
"step": 762
|
|
},
|
|
{
|
|
"epoch": 2.53,
|
|
"learning_rate": 2.4021739130434784e-06,
|
|
"loss": 0.6492,
|
|
"step": 763
|
|
},
|
|
{
|
|
"epoch": 2.53,
|
|
"learning_rate": 2.3967391304347826e-06,
|
|
"loss": 0.6252,
|
|
"step": 764
|
|
},
|
|
{
|
|
"epoch": 2.54,
|
|
"learning_rate": 2.391304347826087e-06,
|
|
"loss": 0.6385,
|
|
"step": 765
|
|
},
|
|
{
|
|
"epoch": 2.54,
|
|
"learning_rate": 2.3858695652173913e-06,
|
|
"loss": 0.6249,
|
|
"step": 766
|
|
},
|
|
{
|
|
"epoch": 2.54,
|
|
"learning_rate": 2.3804347826086955e-06,
|
|
"loss": 0.6272,
|
|
"step": 767
|
|
},
|
|
{
|
|
"epoch": 2.55,
|
|
"learning_rate": 2.375e-06,
|
|
"loss": 0.6533,
|
|
"step": 768
|
|
},
|
|
{
|
|
"epoch": 2.55,
|
|
"learning_rate": 2.3695652173913043e-06,
|
|
"loss": 0.6244,
|
|
"step": 769
|
|
},
|
|
{
|
|
"epoch": 2.55,
|
|
"learning_rate": 2.3641304347826085e-06,
|
|
"loss": 0.6553,
|
|
"step": 770
|
|
},
|
|
{
|
|
"epoch": 2.56,
|
|
"learning_rate": 2.358695652173913e-06,
|
|
"loss": 0.6257,
|
|
"step": 771
|
|
},
|
|
{
|
|
"epoch": 2.56,
|
|
"learning_rate": 2.3532608695652177e-06,
|
|
"loss": 0.6266,
|
|
"step": 772
|
|
},
|
|
{
|
|
"epoch": 2.56,
|
|
"learning_rate": 2.347826086956522e-06,
|
|
"loss": 0.6351,
|
|
"step": 773
|
|
},
|
|
{
|
|
"epoch": 2.57,
|
|
"learning_rate": 2.342391304347826e-06,
|
|
"loss": 0.6399,
|
|
"step": 774
|
|
},
|
|
{
|
|
"epoch": 2.57,
|
|
"learning_rate": 2.3369565217391307e-06,
|
|
"loss": 0.6195,
|
|
"step": 775
|
|
},
|
|
{
|
|
"epoch": 2.57,
|
|
"learning_rate": 2.331521739130435e-06,
|
|
"loss": 0.6235,
|
|
"step": 776
|
|
},
|
|
{
|
|
"epoch": 2.58,
|
|
"learning_rate": 2.326086956521739e-06,
|
|
"loss": 0.6235,
|
|
"step": 777
|
|
},
|
|
{
|
|
"epoch": 2.58,
|
|
"learning_rate": 2.3206521739130436e-06,
|
|
"loss": 0.6143,
|
|
"step": 778
|
|
},
|
|
{
|
|
"epoch": 2.58,
|
|
"learning_rate": 2.315217391304348e-06,
|
|
"loss": 0.6099,
|
|
"step": 779
|
|
},
|
|
{
|
|
"epoch": 2.59,
|
|
"learning_rate": 2.309782608695652e-06,
|
|
"loss": 0.63,
|
|
"step": 780
|
|
},
|
|
{
|
|
"epoch": 2.59,
|
|
"learning_rate": 2.3043478260869566e-06,
|
|
"loss": 0.6349,
|
|
"step": 781
|
|
},
|
|
{
|
|
"epoch": 2.59,
|
|
"learning_rate": 2.2989130434782608e-06,
|
|
"loss": 0.6158,
|
|
"step": 782
|
|
},
|
|
{
|
|
"epoch": 2.6,
|
|
"learning_rate": 2.293478260869565e-06,
|
|
"loss": 0.6386,
|
|
"step": 783
|
|
},
|
|
{
|
|
"epoch": 2.6,
|
|
"learning_rate": 2.2880434782608695e-06,
|
|
"loss": 0.6351,
|
|
"step": 784
|
|
},
|
|
{
|
|
"epoch": 2.6,
|
|
"learning_rate": 2.282608695652174e-06,
|
|
"loss": 0.6264,
|
|
"step": 785
|
|
},
|
|
{
|
|
"epoch": 2.61,
|
|
"learning_rate": 2.2771739130434783e-06,
|
|
"loss": 0.6343,
|
|
"step": 786
|
|
},
|
|
{
|
|
"epoch": 2.61,
|
|
"learning_rate": 2.271739130434783e-06,
|
|
"loss": 0.6395,
|
|
"step": 787
|
|
},
|
|
{
|
|
"epoch": 2.61,
|
|
"learning_rate": 2.266304347826087e-06,
|
|
"loss": 0.6398,
|
|
"step": 788
|
|
},
|
|
{
|
|
"epoch": 2.62,
|
|
"learning_rate": 2.2608695652173913e-06,
|
|
"loss": 0.6329,
|
|
"step": 789
|
|
},
|
|
{
|
|
"epoch": 2.62,
|
|
"learning_rate": 2.255434782608696e-06,
|
|
"loss": 0.6385,
|
|
"step": 790
|
|
},
|
|
{
|
|
"epoch": 2.62,
|
|
"learning_rate": 2.25e-06,
|
|
"loss": 0.6279,
|
|
"step": 791
|
|
},
|
|
{
|
|
"epoch": 2.63,
|
|
"learning_rate": 2.2445652173913042e-06,
|
|
"loss": 0.6235,
|
|
"step": 792
|
|
},
|
|
{
|
|
"epoch": 2.63,
|
|
"learning_rate": 2.239130434782609e-06,
|
|
"loss": 0.6309,
|
|
"step": 793
|
|
},
|
|
{
|
|
"epoch": 2.63,
|
|
"eval_loss": 0.7672982215881348,
|
|
"eval_runtime": 17.6155,
|
|
"eval_samples_per_second": 129.545,
|
|
"eval_steps_per_second": 5.45,
|
|
"step": 793
|
|
},
|
|
{
|
|
"epoch": 2.63,
|
|
"learning_rate": 2.233695652173913e-06,
|
|
"loss": 0.6449,
|
|
"step": 794
|
|
},
|
|
{
|
|
"epoch": 2.64,
|
|
"learning_rate": 2.228260869565217e-06,
|
|
"loss": 0.6438,
|
|
"step": 795
|
|
},
|
|
{
|
|
"epoch": 2.64,
|
|
"learning_rate": 2.222826086956522e-06,
|
|
"loss": 0.6295,
|
|
"step": 796
|
|
},
|
|
{
|
|
"epoch": 2.64,
|
|
"learning_rate": 2.217391304347826e-06,
|
|
"loss": 0.6437,
|
|
"step": 797
|
|
},
|
|
{
|
|
"epoch": 2.65,
|
|
"learning_rate": 2.2119565217391306e-06,
|
|
"loss": 0.6383,
|
|
"step": 798
|
|
},
|
|
{
|
|
"epoch": 2.65,
|
|
"learning_rate": 2.206521739130435e-06,
|
|
"loss": 0.6198,
|
|
"step": 799
|
|
},
|
|
{
|
|
"epoch": 2.65,
|
|
"learning_rate": 2.2010869565217394e-06,
|
|
"loss": 0.635,
|
|
"step": 800
|
|
},
|
|
{
|
|
"epoch": 2.66,
|
|
"learning_rate": 2.1956521739130435e-06,
|
|
"loss": 0.6344,
|
|
"step": 801
|
|
},
|
|
{
|
|
"epoch": 2.66,
|
|
"learning_rate": 2.190217391304348e-06,
|
|
"loss": 0.6427,
|
|
"step": 802
|
|
},
|
|
{
|
|
"epoch": 2.66,
|
|
"learning_rate": 2.1847826086956523e-06,
|
|
"loss": 0.6346,
|
|
"step": 803
|
|
},
|
|
{
|
|
"epoch": 2.67,
|
|
"learning_rate": 2.1793478260869565e-06,
|
|
"loss": 0.6427,
|
|
"step": 804
|
|
},
|
|
{
|
|
"epoch": 2.67,
|
|
"learning_rate": 2.173913043478261e-06,
|
|
"loss": 0.6394,
|
|
"step": 805
|
|
},
|
|
{
|
|
"epoch": 2.67,
|
|
"learning_rate": 2.1684782608695653e-06,
|
|
"loss": 0.6446,
|
|
"step": 806
|
|
},
|
|
{
|
|
"epoch": 2.68,
|
|
"learning_rate": 2.1630434782608695e-06,
|
|
"loss": 0.6453,
|
|
"step": 807
|
|
},
|
|
{
|
|
"epoch": 2.68,
|
|
"learning_rate": 2.157608695652174e-06,
|
|
"loss": 0.6413,
|
|
"step": 808
|
|
},
|
|
{
|
|
"epoch": 2.68,
|
|
"learning_rate": 2.1521739130434782e-06,
|
|
"loss": 0.6522,
|
|
"step": 809
|
|
},
|
|
{
|
|
"epoch": 2.69,
|
|
"learning_rate": 2.1467391304347824e-06,
|
|
"loss": 0.6361,
|
|
"step": 810
|
|
},
|
|
{
|
|
"epoch": 2.69,
|
|
"learning_rate": 2.141304347826087e-06,
|
|
"loss": 0.646,
|
|
"step": 811
|
|
},
|
|
{
|
|
"epoch": 2.69,
|
|
"learning_rate": 2.1358695652173916e-06,
|
|
"loss": 0.6367,
|
|
"step": 812
|
|
},
|
|
{
|
|
"epoch": 2.7,
|
|
"learning_rate": 2.130434782608696e-06,
|
|
"loss": 0.633,
|
|
"step": 813
|
|
},
|
|
{
|
|
"epoch": 2.7,
|
|
"learning_rate": 2.125e-06,
|
|
"loss": 0.6363,
|
|
"step": 814
|
|
},
|
|
{
|
|
"epoch": 2.7,
|
|
"learning_rate": 2.1195652173913046e-06,
|
|
"loss": 0.6244,
|
|
"step": 815
|
|
},
|
|
{
|
|
"epoch": 2.71,
|
|
"learning_rate": 2.1141304347826088e-06,
|
|
"loss": 0.6339,
|
|
"step": 816
|
|
},
|
|
{
|
|
"epoch": 2.71,
|
|
"learning_rate": 2.108695652173913e-06,
|
|
"loss": 0.6496,
|
|
"step": 817
|
|
},
|
|
{
|
|
"epoch": 2.71,
|
|
"learning_rate": 2.1032608695652175e-06,
|
|
"loss": 0.637,
|
|
"step": 818
|
|
},
|
|
{
|
|
"epoch": 2.72,
|
|
"learning_rate": 2.0978260869565217e-06,
|
|
"loss": 0.6239,
|
|
"step": 819
|
|
},
|
|
{
|
|
"epoch": 2.72,
|
|
"learning_rate": 2.092391304347826e-06,
|
|
"loss": 0.6361,
|
|
"step": 820
|
|
},
|
|
{
|
|
"epoch": 2.72,
|
|
"learning_rate": 2.0869565217391305e-06,
|
|
"loss": 0.6395,
|
|
"step": 821
|
|
},
|
|
{
|
|
"epoch": 2.73,
|
|
"learning_rate": 2.0815217391304347e-06,
|
|
"loss": 0.6114,
|
|
"step": 822
|
|
},
|
|
{
|
|
"epoch": 2.73,
|
|
"learning_rate": 2.0760869565217393e-06,
|
|
"loss": 0.6271,
|
|
"step": 823
|
|
},
|
|
{
|
|
"epoch": 2.73,
|
|
"learning_rate": 2.0706521739130435e-06,
|
|
"loss": 0.6286,
|
|
"step": 824
|
|
},
|
|
{
|
|
"epoch": 2.74,
|
|
"learning_rate": 2.0652173913043476e-06,
|
|
"loss": 0.6325,
|
|
"step": 825
|
|
},
|
|
{
|
|
"epoch": 2.74,
|
|
"learning_rate": 2.0597826086956522e-06,
|
|
"loss": 0.6355,
|
|
"step": 826
|
|
},
|
|
{
|
|
"epoch": 2.74,
|
|
"learning_rate": 2.054347826086957e-06,
|
|
"loss": 0.6335,
|
|
"step": 827
|
|
},
|
|
{
|
|
"epoch": 2.75,
|
|
"learning_rate": 2.048913043478261e-06,
|
|
"loss": 0.6422,
|
|
"step": 828
|
|
},
|
|
{
|
|
"epoch": 2.75,
|
|
"learning_rate": 2.043478260869565e-06,
|
|
"loss": 0.6407,
|
|
"step": 829
|
|
},
|
|
{
|
|
"epoch": 2.75,
|
|
"learning_rate": 2.03804347826087e-06,
|
|
"loss": 0.6405,
|
|
"step": 830
|
|
},
|
|
{
|
|
"epoch": 2.76,
|
|
"learning_rate": 2.032608695652174e-06,
|
|
"loss": 0.6409,
|
|
"step": 831
|
|
},
|
|
{
|
|
"epoch": 2.76,
|
|
"learning_rate": 2.027173913043478e-06,
|
|
"loss": 0.6561,
|
|
"step": 832
|
|
},
|
|
{
|
|
"epoch": 2.76,
|
|
"learning_rate": 2.0217391304347828e-06,
|
|
"loss": 0.6346,
|
|
"step": 833
|
|
},
|
|
{
|
|
"epoch": 2.77,
|
|
"learning_rate": 2.016304347826087e-06,
|
|
"loss": 0.6385,
|
|
"step": 834
|
|
},
|
|
{
|
|
"epoch": 2.77,
|
|
"learning_rate": 2.010869565217391e-06,
|
|
"loss": 0.6325,
|
|
"step": 835
|
|
},
|
|
{
|
|
"epoch": 2.77,
|
|
"learning_rate": 2.0054347826086957e-06,
|
|
"loss": 0.6299,
|
|
"step": 836
|
|
},
|
|
{
|
|
"epoch": 2.78,
|
|
"learning_rate": 2e-06,
|
|
"loss": 0.6252,
|
|
"step": 837
|
|
},
|
|
{
|
|
"epoch": 2.78,
|
|
"learning_rate": 1.994565217391304e-06,
|
|
"loss": 0.6418,
|
|
"step": 838
|
|
},
|
|
{
|
|
"epoch": 2.78,
|
|
"learning_rate": 1.9891304347826087e-06,
|
|
"loss": 0.6263,
|
|
"step": 839
|
|
},
|
|
{
|
|
"epoch": 2.79,
|
|
"learning_rate": 1.9836956521739133e-06,
|
|
"loss": 0.6293,
|
|
"step": 840
|
|
},
|
|
{
|
|
"epoch": 2.79,
|
|
"learning_rate": 1.9782608695652175e-06,
|
|
"loss": 0.6288,
|
|
"step": 841
|
|
},
|
|
{
|
|
"epoch": 2.79,
|
|
"learning_rate": 1.972826086956522e-06,
|
|
"loss": 0.6351,
|
|
"step": 842
|
|
},
|
|
{
|
|
"epoch": 2.8,
|
|
"learning_rate": 1.9673913043478263e-06,
|
|
"loss": 0.6285,
|
|
"step": 843
|
|
},
|
|
{
|
|
"epoch": 2.8,
|
|
"learning_rate": 1.9619565217391304e-06,
|
|
"loss": 0.6307,
|
|
"step": 844
|
|
},
|
|
{
|
|
"epoch": 2.8,
|
|
"learning_rate": 1.956521739130435e-06,
|
|
"loss": 0.6301,
|
|
"step": 845
|
|
},
|
|
{
|
|
"epoch": 2.81,
|
|
"learning_rate": 1.9510869565217392e-06,
|
|
"loss": 0.6226,
|
|
"step": 846
|
|
},
|
|
{
|
|
"epoch": 2.81,
|
|
"learning_rate": 1.9456521739130434e-06,
|
|
"loss": 0.622,
|
|
"step": 847
|
|
},
|
|
{
|
|
"epoch": 2.81,
|
|
"learning_rate": 1.940217391304348e-06,
|
|
"loss": 0.6267,
|
|
"step": 848
|
|
},
|
|
{
|
|
"epoch": 2.82,
|
|
"learning_rate": 1.934782608695652e-06,
|
|
"loss": 0.6106,
|
|
"step": 849
|
|
},
|
|
{
|
|
"epoch": 2.82,
|
|
"learning_rate": 1.9293478260869564e-06,
|
|
"loss": 0.6415,
|
|
"step": 850
|
|
},
|
|
{
|
|
"epoch": 2.82,
|
|
"learning_rate": 1.923913043478261e-06,
|
|
"loss": 0.6326,
|
|
"step": 851
|
|
},
|
|
{
|
|
"epoch": 2.83,
|
|
"learning_rate": 1.918478260869565e-06,
|
|
"loss": 0.6341,
|
|
"step": 852
|
|
},
|
|
{
|
|
"epoch": 2.83,
|
|
"learning_rate": 1.9130434782608693e-06,
|
|
"loss": 0.644,
|
|
"step": 853
|
|
},
|
|
{
|
|
"epoch": 2.83,
|
|
"learning_rate": 1.907608695652174e-06,
|
|
"loss": 0.6255,
|
|
"step": 854
|
|
},
|
|
{
|
|
"epoch": 2.83,
|
|
"eval_loss": 0.7656497359275818,
|
|
"eval_runtime": 17.6221,
|
|
"eval_samples_per_second": 129.496,
|
|
"eval_steps_per_second": 5.448,
|
|
"step": 854
|
|
},
|
|
{
|
|
"epoch": 2.84,
|
|
"learning_rate": 1.9021739130434785e-06,
|
|
"loss": 0.6199,
|
|
"step": 855
|
|
},
|
|
{
|
|
"epoch": 2.84,
|
|
"learning_rate": 1.8967391304347827e-06,
|
|
"loss": 0.6334,
|
|
"step": 856
|
|
},
|
|
{
|
|
"epoch": 2.84,
|
|
"learning_rate": 1.891304347826087e-06,
|
|
"loss": 0.6348,
|
|
"step": 857
|
|
},
|
|
{
|
|
"epoch": 2.85,
|
|
"learning_rate": 1.8858695652173915e-06,
|
|
"loss": 0.6361,
|
|
"step": 858
|
|
},
|
|
{
|
|
"epoch": 2.85,
|
|
"learning_rate": 1.8804347826086957e-06,
|
|
"loss": 0.6388,
|
|
"step": 859
|
|
},
|
|
{
|
|
"epoch": 2.85,
|
|
"learning_rate": 1.875e-06,
|
|
"loss": 0.653,
|
|
"step": 860
|
|
},
|
|
{
|
|
"epoch": 2.86,
|
|
"learning_rate": 1.8695652173913044e-06,
|
|
"loss": 0.6391,
|
|
"step": 861
|
|
},
|
|
{
|
|
"epoch": 2.86,
|
|
"learning_rate": 1.8641304347826086e-06,
|
|
"loss": 0.6152,
|
|
"step": 862
|
|
},
|
|
{
|
|
"epoch": 2.86,
|
|
"learning_rate": 1.858695652173913e-06,
|
|
"loss": 0.6552,
|
|
"step": 863
|
|
},
|
|
{
|
|
"epoch": 2.87,
|
|
"learning_rate": 1.8532608695652174e-06,
|
|
"loss": 0.6354,
|
|
"step": 864
|
|
},
|
|
{
|
|
"epoch": 2.87,
|
|
"learning_rate": 1.8478260869565216e-06,
|
|
"loss": 0.6272,
|
|
"step": 865
|
|
},
|
|
{
|
|
"epoch": 2.87,
|
|
"learning_rate": 1.842391304347826e-06,
|
|
"loss": 0.6433,
|
|
"step": 866
|
|
},
|
|
{
|
|
"epoch": 2.88,
|
|
"learning_rate": 1.8369565217391306e-06,
|
|
"loss": 0.6174,
|
|
"step": 867
|
|
},
|
|
{
|
|
"epoch": 2.88,
|
|
"learning_rate": 1.831521739130435e-06,
|
|
"loss": 0.6129,
|
|
"step": 868
|
|
},
|
|
{
|
|
"epoch": 2.88,
|
|
"learning_rate": 1.8260869565217394e-06,
|
|
"loss": 0.6384,
|
|
"step": 869
|
|
},
|
|
{
|
|
"epoch": 2.89,
|
|
"learning_rate": 1.8206521739130435e-06,
|
|
"loss": 0.6471,
|
|
"step": 870
|
|
},
|
|
{
|
|
"epoch": 2.89,
|
|
"learning_rate": 1.815217391304348e-06,
|
|
"loss": 0.6339,
|
|
"step": 871
|
|
},
|
|
{
|
|
"epoch": 2.89,
|
|
"learning_rate": 1.8097826086956523e-06,
|
|
"loss": 0.6236,
|
|
"step": 872
|
|
},
|
|
{
|
|
"epoch": 2.9,
|
|
"learning_rate": 1.8043478260869565e-06,
|
|
"loss": 0.6083,
|
|
"step": 873
|
|
},
|
|
{
|
|
"epoch": 2.9,
|
|
"learning_rate": 1.7989130434782609e-06,
|
|
"loss": 0.6331,
|
|
"step": 874
|
|
},
|
|
{
|
|
"epoch": 2.9,
|
|
"learning_rate": 1.7934782608695653e-06,
|
|
"loss": 0.634,
|
|
"step": 875
|
|
},
|
|
{
|
|
"epoch": 2.91,
|
|
"learning_rate": 1.7880434782608697e-06,
|
|
"loss": 0.63,
|
|
"step": 876
|
|
},
|
|
{
|
|
"epoch": 2.91,
|
|
"learning_rate": 1.7826086956521738e-06,
|
|
"loss": 0.6643,
|
|
"step": 877
|
|
},
|
|
{
|
|
"epoch": 2.91,
|
|
"learning_rate": 1.7771739130434782e-06,
|
|
"loss": 0.639,
|
|
"step": 878
|
|
},
|
|
{
|
|
"epoch": 2.92,
|
|
"learning_rate": 1.7717391304347826e-06,
|
|
"loss": 0.6396,
|
|
"step": 879
|
|
},
|
|
{
|
|
"epoch": 2.92,
|
|
"learning_rate": 1.7663043478260868e-06,
|
|
"loss": 0.6391,
|
|
"step": 880
|
|
},
|
|
{
|
|
"epoch": 2.92,
|
|
"learning_rate": 1.7608695652173914e-06,
|
|
"loss": 0.6264,
|
|
"step": 881
|
|
},
|
|
{
|
|
"epoch": 2.93,
|
|
"learning_rate": 1.7554347826086958e-06,
|
|
"loss": 0.6211,
|
|
"step": 882
|
|
},
|
|
{
|
|
"epoch": 2.93,
|
|
"learning_rate": 1.7500000000000002e-06,
|
|
"loss": 0.6119,
|
|
"step": 883
|
|
},
|
|
{
|
|
"epoch": 2.93,
|
|
"learning_rate": 1.7445652173913046e-06,
|
|
"loss": 0.6365,
|
|
"step": 884
|
|
},
|
|
{
|
|
"epoch": 2.94,
|
|
"learning_rate": 1.7391304347826088e-06,
|
|
"loss": 0.6396,
|
|
"step": 885
|
|
},
|
|
{
|
|
"epoch": 2.94,
|
|
"learning_rate": 1.7336956521739131e-06,
|
|
"loss": 0.6489,
|
|
"step": 886
|
|
},
|
|
{
|
|
"epoch": 2.94,
|
|
"learning_rate": 1.7282608695652175e-06,
|
|
"loss": 0.6471,
|
|
"step": 887
|
|
},
|
|
{
|
|
"epoch": 2.95,
|
|
"learning_rate": 1.7228260869565217e-06,
|
|
"loss": 0.6243,
|
|
"step": 888
|
|
},
|
|
{
|
|
"epoch": 2.95,
|
|
"learning_rate": 1.7173913043478261e-06,
|
|
"loss": 0.6238,
|
|
"step": 889
|
|
},
|
|
{
|
|
"epoch": 2.95,
|
|
"learning_rate": 1.7119565217391305e-06,
|
|
"loss": 0.6162,
|
|
"step": 890
|
|
},
|
|
{
|
|
"epoch": 2.96,
|
|
"learning_rate": 1.7065217391304347e-06,
|
|
"loss": 0.6443,
|
|
"step": 891
|
|
},
|
|
{
|
|
"epoch": 2.96,
|
|
"learning_rate": 1.701086956521739e-06,
|
|
"loss": 0.637,
|
|
"step": 892
|
|
},
|
|
{
|
|
"epoch": 2.96,
|
|
"learning_rate": 1.6956521739130435e-06,
|
|
"loss": 0.6377,
|
|
"step": 893
|
|
},
|
|
{
|
|
"epoch": 2.97,
|
|
"learning_rate": 1.6902173913043476e-06,
|
|
"loss": 0.6316,
|
|
"step": 894
|
|
},
|
|
{
|
|
"epoch": 2.97,
|
|
"learning_rate": 1.6847826086956524e-06,
|
|
"loss": 0.6405,
|
|
"step": 895
|
|
},
|
|
{
|
|
"epoch": 2.97,
|
|
"learning_rate": 1.6793478260869566e-06,
|
|
"loss": 0.6483,
|
|
"step": 896
|
|
},
|
|
{
|
|
"epoch": 2.98,
|
|
"learning_rate": 1.673913043478261e-06,
|
|
"loss": 0.6318,
|
|
"step": 897
|
|
},
|
|
{
|
|
"epoch": 2.98,
|
|
"learning_rate": 1.6684782608695654e-06,
|
|
"loss": 0.6447,
|
|
"step": 898
|
|
},
|
|
{
|
|
"epoch": 2.98,
|
|
"learning_rate": 1.6630434782608696e-06,
|
|
"loss": 0.65,
|
|
"step": 899
|
|
},
|
|
{
|
|
"epoch": 2.99,
|
|
"learning_rate": 1.657608695652174e-06,
|
|
"loss": 0.6443,
|
|
"step": 900
|
|
},
|
|
{
|
|
"epoch": 2.99,
|
|
"learning_rate": 1.6521739130434784e-06,
|
|
"loss": 0.6355,
|
|
"step": 901
|
|
},
|
|
{
|
|
"epoch": 2.99,
|
|
"learning_rate": 1.6467391304347825e-06,
|
|
"loss": 0.642,
|
|
"step": 902
|
|
},
|
|
{
|
|
"epoch": 3.0,
|
|
"learning_rate": 1.641304347826087e-06,
|
|
"loss": 0.6527,
|
|
"step": 903
|
|
},
|
|
{
|
|
"epoch": 3.0,
|
|
"learning_rate": 1.6358695652173913e-06,
|
|
"loss": 0.6383,
|
|
"step": 904
|
|
},
|
|
{
|
|
"epoch": 3.0,
|
|
"learning_rate": 1.6304347826086955e-06,
|
|
"loss": 0.6181,
|
|
"step": 905
|
|
},
|
|
{
|
|
"epoch": 3.0,
|
|
"learning_rate": 1.625e-06,
|
|
"loss": 0.5692,
|
|
"step": 906
|
|
},
|
|
{
|
|
"epoch": 3.01,
|
|
"learning_rate": 1.6195652173913043e-06,
|
|
"loss": 0.5856,
|
|
"step": 907
|
|
},
|
|
{
|
|
"epoch": 3.01,
|
|
"learning_rate": 1.6141304347826087e-06,
|
|
"loss": 0.5873,
|
|
"step": 908
|
|
},
|
|
{
|
|
"epoch": 3.01,
|
|
"learning_rate": 1.6086956521739133e-06,
|
|
"loss": 0.5891,
|
|
"step": 909
|
|
},
|
|
{
|
|
"epoch": 3.02,
|
|
"learning_rate": 1.6032608695652175e-06,
|
|
"loss": 0.5623,
|
|
"step": 910
|
|
},
|
|
{
|
|
"epoch": 3.02,
|
|
"learning_rate": 1.5978260869565219e-06,
|
|
"loss": 0.6019,
|
|
"step": 911
|
|
},
|
|
{
|
|
"epoch": 3.02,
|
|
"learning_rate": 1.5923913043478262e-06,
|
|
"loss": 0.5961,
|
|
"step": 912
|
|
},
|
|
{
|
|
"epoch": 3.03,
|
|
"learning_rate": 1.5869565217391304e-06,
|
|
"loss": 0.5779,
|
|
"step": 913
|
|
},
|
|
{
|
|
"epoch": 3.03,
|
|
"learning_rate": 1.5815217391304348e-06,
|
|
"loss": 0.5728,
|
|
"step": 914
|
|
},
|
|
{
|
|
"epoch": 3.03,
|
|
"learning_rate": 1.5760869565217392e-06,
|
|
"loss": 0.5855,
|
|
"step": 915
|
|
},
|
|
{
|
|
"epoch": 3.03,
|
|
"eval_loss": 0.7823443412780762,
|
|
"eval_runtime": 17.617,
|
|
"eval_samples_per_second": 129.534,
|
|
"eval_steps_per_second": 5.449,
|
|
"step": 915
|
|
},
|
|
{
|
|
"epoch": 3.04,
|
|
"learning_rate": 1.5706521739130436e-06,
|
|
"loss": 0.5867,
|
|
"step": 916
|
|
},
|
|
{
|
|
"epoch": 3.04,
|
|
"learning_rate": 1.5652173913043478e-06,
|
|
"loss": 0.5808,
|
|
"step": 917
|
|
},
|
|
{
|
|
"epoch": 3.04,
|
|
"learning_rate": 1.5597826086956522e-06,
|
|
"loss": 0.5879,
|
|
"step": 918
|
|
},
|
|
{
|
|
"epoch": 3.05,
|
|
"learning_rate": 1.5543478260869566e-06,
|
|
"loss": 0.5726,
|
|
"step": 919
|
|
},
|
|
{
|
|
"epoch": 3.05,
|
|
"learning_rate": 1.5489130434782607e-06,
|
|
"loss": 0.5633,
|
|
"step": 920
|
|
},
|
|
{
|
|
"epoch": 3.05,
|
|
"learning_rate": 1.5434782608695651e-06,
|
|
"loss": 0.5703,
|
|
"step": 921
|
|
},
|
|
{
|
|
"epoch": 3.06,
|
|
"learning_rate": 1.5380434782608695e-06,
|
|
"loss": 0.5811,
|
|
"step": 922
|
|
},
|
|
{
|
|
"epoch": 3.06,
|
|
"learning_rate": 1.5326086956521741e-06,
|
|
"loss": 0.5724,
|
|
"step": 923
|
|
},
|
|
{
|
|
"epoch": 3.06,
|
|
"learning_rate": 1.5271739130434785e-06,
|
|
"loss": 0.5885,
|
|
"step": 924
|
|
},
|
|
{
|
|
"epoch": 3.07,
|
|
"learning_rate": 1.5217391304347827e-06,
|
|
"loss": 0.5744,
|
|
"step": 925
|
|
},
|
|
{
|
|
"epoch": 3.07,
|
|
"learning_rate": 1.516304347826087e-06,
|
|
"loss": 0.5658,
|
|
"step": 926
|
|
},
|
|
{
|
|
"epoch": 3.07,
|
|
"learning_rate": 1.5108695652173915e-06,
|
|
"loss": 0.5684,
|
|
"step": 927
|
|
},
|
|
{
|
|
"epoch": 3.08,
|
|
"learning_rate": 1.5054347826086956e-06,
|
|
"loss": 0.5803,
|
|
"step": 928
|
|
},
|
|
{
|
|
"epoch": 3.08,
|
|
"learning_rate": 1.5e-06,
|
|
"loss": 0.586,
|
|
"step": 929
|
|
},
|
|
{
|
|
"epoch": 3.08,
|
|
"learning_rate": 1.4945652173913044e-06,
|
|
"loss": 0.5787,
|
|
"step": 930
|
|
},
|
|
{
|
|
"epoch": 3.09,
|
|
"learning_rate": 1.4891304347826086e-06,
|
|
"loss": 0.5757,
|
|
"step": 931
|
|
},
|
|
{
|
|
"epoch": 3.09,
|
|
"learning_rate": 1.483695652173913e-06,
|
|
"loss": 0.5655,
|
|
"step": 932
|
|
},
|
|
{
|
|
"epoch": 3.09,
|
|
"learning_rate": 1.4782608695652176e-06,
|
|
"loss": 0.5827,
|
|
"step": 933
|
|
},
|
|
{
|
|
"epoch": 3.1,
|
|
"learning_rate": 1.4728260869565218e-06,
|
|
"loss": 0.5978,
|
|
"step": 934
|
|
},
|
|
{
|
|
"epoch": 3.1,
|
|
"learning_rate": 1.4673913043478262e-06,
|
|
"loss": 0.5797,
|
|
"step": 935
|
|
},
|
|
{
|
|
"epoch": 3.1,
|
|
"learning_rate": 1.4619565217391306e-06,
|
|
"loss": 0.5823,
|
|
"step": 936
|
|
},
|
|
{
|
|
"epoch": 3.11,
|
|
"learning_rate": 1.4565217391304347e-06,
|
|
"loss": 0.5805,
|
|
"step": 937
|
|
},
|
|
{
|
|
"epoch": 3.11,
|
|
"learning_rate": 1.4510869565217391e-06,
|
|
"loss": 0.5685,
|
|
"step": 938
|
|
},
|
|
{
|
|
"epoch": 3.11,
|
|
"learning_rate": 1.4456521739130435e-06,
|
|
"loss": 0.5818,
|
|
"step": 939
|
|
},
|
|
{
|
|
"epoch": 3.12,
|
|
"learning_rate": 1.440217391304348e-06,
|
|
"loss": 0.571,
|
|
"step": 940
|
|
},
|
|
{
|
|
"epoch": 3.12,
|
|
"learning_rate": 1.4347826086956523e-06,
|
|
"loss": 0.5962,
|
|
"step": 941
|
|
},
|
|
{
|
|
"epoch": 3.12,
|
|
"learning_rate": 1.4293478260869565e-06,
|
|
"loss": 0.5897,
|
|
"step": 942
|
|
},
|
|
{
|
|
"epoch": 3.13,
|
|
"learning_rate": 1.4239130434782609e-06,
|
|
"loss": 0.5979,
|
|
"step": 943
|
|
},
|
|
{
|
|
"epoch": 3.13,
|
|
"learning_rate": 1.4184782608695653e-06,
|
|
"loss": 0.5771,
|
|
"step": 944
|
|
},
|
|
{
|
|
"epoch": 3.13,
|
|
"learning_rate": 1.4130434782608697e-06,
|
|
"loss": 0.5832,
|
|
"step": 945
|
|
},
|
|
{
|
|
"epoch": 3.14,
|
|
"learning_rate": 1.4076086956521738e-06,
|
|
"loss": 0.5997,
|
|
"step": 946
|
|
},
|
|
{
|
|
"epoch": 3.14,
|
|
"learning_rate": 1.4021739130434784e-06,
|
|
"loss": 0.5826,
|
|
"step": 947
|
|
},
|
|
{
|
|
"epoch": 3.14,
|
|
"learning_rate": 1.3967391304347826e-06,
|
|
"loss": 0.5864,
|
|
"step": 948
|
|
},
|
|
{
|
|
"epoch": 3.15,
|
|
"learning_rate": 1.391304347826087e-06,
|
|
"loss": 0.5822,
|
|
"step": 949
|
|
},
|
|
{
|
|
"epoch": 3.15,
|
|
"learning_rate": 1.3858695652173914e-06,
|
|
"loss": 0.5948,
|
|
"step": 950
|
|
},
|
|
{
|
|
"epoch": 3.15,
|
|
"learning_rate": 1.3804347826086956e-06,
|
|
"loss": 0.5913,
|
|
"step": 951
|
|
},
|
|
{
|
|
"epoch": 3.16,
|
|
"learning_rate": 1.375e-06,
|
|
"loss": 0.5886,
|
|
"step": 952
|
|
},
|
|
{
|
|
"epoch": 3.16,
|
|
"learning_rate": 1.3695652173913044e-06,
|
|
"loss": 0.5773,
|
|
"step": 953
|
|
},
|
|
{
|
|
"epoch": 3.16,
|
|
"learning_rate": 1.3641304347826087e-06,
|
|
"loss": 0.6007,
|
|
"step": 954
|
|
},
|
|
{
|
|
"epoch": 3.17,
|
|
"learning_rate": 1.3586956521739131e-06,
|
|
"loss": 0.5644,
|
|
"step": 955
|
|
},
|
|
{
|
|
"epoch": 3.17,
|
|
"learning_rate": 1.3532608695652175e-06,
|
|
"loss": 0.6069,
|
|
"step": 956
|
|
},
|
|
{
|
|
"epoch": 3.17,
|
|
"learning_rate": 1.3478260869565217e-06,
|
|
"loss": 0.5777,
|
|
"step": 957
|
|
},
|
|
{
|
|
"epoch": 3.18,
|
|
"learning_rate": 1.342391304347826e-06,
|
|
"loss": 0.5901,
|
|
"step": 958
|
|
},
|
|
{
|
|
"epoch": 3.18,
|
|
"learning_rate": 1.3369565217391305e-06,
|
|
"loss": 0.5853,
|
|
"step": 959
|
|
},
|
|
{
|
|
"epoch": 3.18,
|
|
"learning_rate": 1.3315217391304347e-06,
|
|
"loss": 0.569,
|
|
"step": 960
|
|
},
|
|
{
|
|
"epoch": 3.19,
|
|
"learning_rate": 1.3260869565217393e-06,
|
|
"loss": 0.5809,
|
|
"step": 961
|
|
},
|
|
{
|
|
"epoch": 3.19,
|
|
"learning_rate": 1.3206521739130434e-06,
|
|
"loss": 0.6015,
|
|
"step": 962
|
|
},
|
|
{
|
|
"epoch": 3.19,
|
|
"learning_rate": 1.3152173913043478e-06,
|
|
"loss": 0.5966,
|
|
"step": 963
|
|
},
|
|
{
|
|
"epoch": 3.2,
|
|
"learning_rate": 1.3097826086956522e-06,
|
|
"loss": 0.5774,
|
|
"step": 964
|
|
},
|
|
{
|
|
"epoch": 3.2,
|
|
"learning_rate": 1.3043478260869566e-06,
|
|
"loss": 0.5777,
|
|
"step": 965
|
|
},
|
|
{
|
|
"epoch": 3.2,
|
|
"learning_rate": 1.2989130434782608e-06,
|
|
"loss": 0.5811,
|
|
"step": 966
|
|
},
|
|
{
|
|
"epoch": 3.21,
|
|
"learning_rate": 1.2934782608695654e-06,
|
|
"loss": 0.5633,
|
|
"step": 967
|
|
},
|
|
{
|
|
"epoch": 3.21,
|
|
"learning_rate": 1.2880434782608696e-06,
|
|
"loss": 0.5904,
|
|
"step": 968
|
|
},
|
|
{
|
|
"epoch": 3.21,
|
|
"learning_rate": 1.282608695652174e-06,
|
|
"loss": 0.5732,
|
|
"step": 969
|
|
},
|
|
{
|
|
"epoch": 3.22,
|
|
"learning_rate": 1.2771739130434784e-06,
|
|
"loss": 0.5988,
|
|
"step": 970
|
|
},
|
|
{
|
|
"epoch": 3.22,
|
|
"learning_rate": 1.2717391304347825e-06,
|
|
"loss": 0.5784,
|
|
"step": 971
|
|
},
|
|
{
|
|
"epoch": 3.22,
|
|
"learning_rate": 1.266304347826087e-06,
|
|
"loss": 0.5813,
|
|
"step": 972
|
|
},
|
|
{
|
|
"epoch": 3.23,
|
|
"learning_rate": 1.2608695652173913e-06,
|
|
"loss": 0.5823,
|
|
"step": 973
|
|
},
|
|
{
|
|
"epoch": 3.23,
|
|
"learning_rate": 1.2554347826086957e-06,
|
|
"loss": 0.5879,
|
|
"step": 974
|
|
},
|
|
{
|
|
"epoch": 3.23,
|
|
"learning_rate": 1.25e-06,
|
|
"loss": 0.5847,
|
|
"step": 975
|
|
},
|
|
{
|
|
"epoch": 3.24,
|
|
"learning_rate": 1.2445652173913045e-06,
|
|
"loss": 0.594,
|
|
"step": 976
|
|
},
|
|
{
|
|
"epoch": 3.24,
|
|
"eval_loss": 0.7821070551872253,
|
|
"eval_runtime": 17.6099,
|
|
"eval_samples_per_second": 129.586,
|
|
"eval_steps_per_second": 5.451,
|
|
"step": 976
|
|
},
|
|
{
|
|
"epoch": 3.24,
|
|
"learning_rate": 1.2391304347826087e-06,
|
|
"loss": 0.5858,
|
|
"step": 977
|
|
},
|
|
{
|
|
"epoch": 3.24,
|
|
"learning_rate": 1.233695652173913e-06,
|
|
"loss": 0.5658,
|
|
"step": 978
|
|
},
|
|
{
|
|
"epoch": 3.25,
|
|
"learning_rate": 1.2282608695652175e-06,
|
|
"loss": 0.6045,
|
|
"step": 979
|
|
},
|
|
{
|
|
"epoch": 3.25,
|
|
"learning_rate": 1.2228260869565216e-06,
|
|
"loss": 0.5956,
|
|
"step": 980
|
|
},
|
|
{
|
|
"epoch": 3.25,
|
|
"learning_rate": 1.2173913043478262e-06,
|
|
"loss": 0.5875,
|
|
"step": 981
|
|
},
|
|
{
|
|
"epoch": 3.26,
|
|
"learning_rate": 1.2119565217391304e-06,
|
|
"loss": 0.5927,
|
|
"step": 982
|
|
},
|
|
{
|
|
"epoch": 3.26,
|
|
"learning_rate": 1.2065217391304348e-06,
|
|
"loss": 0.595,
|
|
"step": 983
|
|
},
|
|
{
|
|
"epoch": 3.26,
|
|
"learning_rate": 1.2010869565217392e-06,
|
|
"loss": 0.6002,
|
|
"step": 984
|
|
},
|
|
{
|
|
"epoch": 3.27,
|
|
"learning_rate": 1.1956521739130436e-06,
|
|
"loss": 0.5785,
|
|
"step": 985
|
|
},
|
|
{
|
|
"epoch": 3.27,
|
|
"learning_rate": 1.1902173913043478e-06,
|
|
"loss": 0.5844,
|
|
"step": 986
|
|
},
|
|
{
|
|
"epoch": 3.27,
|
|
"learning_rate": 1.1847826086956522e-06,
|
|
"loss": 0.5722,
|
|
"step": 987
|
|
},
|
|
{
|
|
"epoch": 3.28,
|
|
"learning_rate": 1.1793478260869565e-06,
|
|
"loss": 0.6025,
|
|
"step": 988
|
|
},
|
|
{
|
|
"epoch": 3.28,
|
|
"learning_rate": 1.173913043478261e-06,
|
|
"loss": 0.5794,
|
|
"step": 989
|
|
},
|
|
{
|
|
"epoch": 3.28,
|
|
"learning_rate": 1.1684782608695653e-06,
|
|
"loss": 0.5758,
|
|
"step": 990
|
|
},
|
|
{
|
|
"epoch": 3.29,
|
|
"learning_rate": 1.1630434782608695e-06,
|
|
"loss": 0.5911,
|
|
"step": 991
|
|
},
|
|
{
|
|
"epoch": 3.29,
|
|
"learning_rate": 1.157608695652174e-06,
|
|
"loss": 0.5862,
|
|
"step": 992
|
|
},
|
|
{
|
|
"epoch": 3.29,
|
|
"learning_rate": 1.1521739130434783e-06,
|
|
"loss": 0.5885,
|
|
"step": 993
|
|
},
|
|
{
|
|
"epoch": 3.3,
|
|
"learning_rate": 1.1467391304347825e-06,
|
|
"loss": 0.5874,
|
|
"step": 994
|
|
},
|
|
{
|
|
"epoch": 3.3,
|
|
"learning_rate": 1.141304347826087e-06,
|
|
"loss": 0.5796,
|
|
"step": 995
|
|
},
|
|
{
|
|
"epoch": 3.3,
|
|
"learning_rate": 1.1358695652173915e-06,
|
|
"loss": 0.598,
|
|
"step": 996
|
|
},
|
|
{
|
|
"epoch": 3.31,
|
|
"learning_rate": 1.1304347826086956e-06,
|
|
"loss": 0.5649,
|
|
"step": 997
|
|
},
|
|
{
|
|
"epoch": 3.31,
|
|
"learning_rate": 1.125e-06,
|
|
"loss": 0.5958,
|
|
"step": 998
|
|
},
|
|
{
|
|
"epoch": 3.31,
|
|
"learning_rate": 1.1195652173913044e-06,
|
|
"loss": 0.5721,
|
|
"step": 999
|
|
},
|
|
{
|
|
"epoch": 3.32,
|
|
"learning_rate": 1.1141304347826086e-06,
|
|
"loss": 0.5807,
|
|
"step": 1000
|
|
},
|
|
{
|
|
"epoch": 3.32,
|
|
"learning_rate": 1.108695652173913e-06,
|
|
"loss": 0.5779,
|
|
"step": 1001
|
|
},
|
|
{
|
|
"epoch": 3.32,
|
|
"learning_rate": 1.1032608695652176e-06,
|
|
"loss": 0.5884,
|
|
"step": 1002
|
|
},
|
|
{
|
|
"epoch": 3.33,
|
|
"learning_rate": 1.0978260869565218e-06,
|
|
"loss": 0.5588,
|
|
"step": 1003
|
|
},
|
|
{
|
|
"epoch": 3.33,
|
|
"learning_rate": 1.0923913043478262e-06,
|
|
"loss": 0.587,
|
|
"step": 1004
|
|
},
|
|
{
|
|
"epoch": 3.33,
|
|
"learning_rate": 1.0869565217391306e-06,
|
|
"loss": 0.5947,
|
|
"step": 1005
|
|
},
|
|
{
|
|
"epoch": 3.34,
|
|
"learning_rate": 1.0815217391304347e-06,
|
|
"loss": 0.5878,
|
|
"step": 1006
|
|
},
|
|
{
|
|
"epoch": 3.34,
|
|
"learning_rate": 1.0760869565217391e-06,
|
|
"loss": 0.5799,
|
|
"step": 1007
|
|
},
|
|
{
|
|
"epoch": 3.34,
|
|
"learning_rate": 1.0706521739130435e-06,
|
|
"loss": 0.5883,
|
|
"step": 1008
|
|
},
|
|
{
|
|
"epoch": 3.35,
|
|
"learning_rate": 1.065217391304348e-06,
|
|
"loss": 0.5812,
|
|
"step": 1009
|
|
},
|
|
{
|
|
"epoch": 3.35,
|
|
"learning_rate": 1.0597826086956523e-06,
|
|
"loss": 0.6035,
|
|
"step": 1010
|
|
},
|
|
{
|
|
"epoch": 3.35,
|
|
"learning_rate": 1.0543478260869565e-06,
|
|
"loss": 0.5917,
|
|
"step": 1011
|
|
},
|
|
{
|
|
"epoch": 3.36,
|
|
"learning_rate": 1.0489130434782609e-06,
|
|
"loss": 0.5717,
|
|
"step": 1012
|
|
},
|
|
{
|
|
"epoch": 3.36,
|
|
"learning_rate": 1.0434782608695653e-06,
|
|
"loss": 0.5666,
|
|
"step": 1013
|
|
},
|
|
{
|
|
"epoch": 3.36,
|
|
"learning_rate": 1.0380434782608696e-06,
|
|
"loss": 0.5983,
|
|
"step": 1014
|
|
},
|
|
{
|
|
"epoch": 3.37,
|
|
"learning_rate": 1.0326086956521738e-06,
|
|
"loss": 0.5921,
|
|
"step": 1015
|
|
},
|
|
{
|
|
"epoch": 3.37,
|
|
"learning_rate": 1.0271739130434784e-06,
|
|
"loss": 0.5959,
|
|
"step": 1016
|
|
},
|
|
{
|
|
"epoch": 3.37,
|
|
"learning_rate": 1.0217391304347826e-06,
|
|
"loss": 0.57,
|
|
"step": 1017
|
|
},
|
|
{
|
|
"epoch": 3.38,
|
|
"learning_rate": 1.016304347826087e-06,
|
|
"loss": 0.5806,
|
|
"step": 1018
|
|
},
|
|
{
|
|
"epoch": 3.38,
|
|
"learning_rate": 1.0108695652173914e-06,
|
|
"loss": 0.5865,
|
|
"step": 1019
|
|
},
|
|
{
|
|
"epoch": 3.38,
|
|
"learning_rate": 1.0054347826086956e-06,
|
|
"loss": 0.5943,
|
|
"step": 1020
|
|
},
|
|
{
|
|
"epoch": 3.39,
|
|
"learning_rate": 1e-06,
|
|
"loss": 0.5695,
|
|
"step": 1021
|
|
},
|
|
{
|
|
"epoch": 3.39,
|
|
"learning_rate": 9.945652173913043e-07,
|
|
"loss": 0.5857,
|
|
"step": 1022
|
|
},
|
|
{
|
|
"epoch": 3.39,
|
|
"learning_rate": 9.891304347826087e-07,
|
|
"loss": 0.5908,
|
|
"step": 1023
|
|
},
|
|
{
|
|
"epoch": 3.4,
|
|
"learning_rate": 9.836956521739131e-07,
|
|
"loss": 0.5626,
|
|
"step": 1024
|
|
},
|
|
{
|
|
"epoch": 3.4,
|
|
"learning_rate": 9.782608695652175e-07,
|
|
"loss": 0.5769,
|
|
"step": 1025
|
|
},
|
|
{
|
|
"epoch": 3.4,
|
|
"learning_rate": 9.728260869565217e-07,
|
|
"loss": 0.5884,
|
|
"step": 1026
|
|
},
|
|
{
|
|
"epoch": 3.41,
|
|
"learning_rate": 9.67391304347826e-07,
|
|
"loss": 0.5691,
|
|
"step": 1027
|
|
},
|
|
{
|
|
"epoch": 3.41,
|
|
"learning_rate": 9.619565217391305e-07,
|
|
"loss": 0.5825,
|
|
"step": 1028
|
|
},
|
|
{
|
|
"epoch": 3.41,
|
|
"learning_rate": 9.565217391304347e-07,
|
|
"loss": 0.5812,
|
|
"step": 1029
|
|
},
|
|
{
|
|
"epoch": 3.42,
|
|
"learning_rate": 9.510869565217393e-07,
|
|
"loss": 0.5674,
|
|
"step": 1030
|
|
},
|
|
{
|
|
"epoch": 3.42,
|
|
"learning_rate": 9.456521739130435e-07,
|
|
"loss": 0.5725,
|
|
"step": 1031
|
|
},
|
|
{
|
|
"epoch": 3.42,
|
|
"learning_rate": 9.402173913043478e-07,
|
|
"loss": 0.5788,
|
|
"step": 1032
|
|
},
|
|
{
|
|
"epoch": 3.43,
|
|
"learning_rate": 9.347826086956522e-07,
|
|
"loss": 0.6002,
|
|
"step": 1033
|
|
},
|
|
{
|
|
"epoch": 3.43,
|
|
"learning_rate": 9.293478260869565e-07,
|
|
"loss": 0.5616,
|
|
"step": 1034
|
|
},
|
|
{
|
|
"epoch": 3.43,
|
|
"learning_rate": 9.239130434782608e-07,
|
|
"loss": 0.5866,
|
|
"step": 1035
|
|
},
|
|
{
|
|
"epoch": 3.44,
|
|
"learning_rate": 9.184782608695653e-07,
|
|
"loss": 0.5993,
|
|
"step": 1036
|
|
},
|
|
{
|
|
"epoch": 3.44,
|
|
"learning_rate": 9.130434782608697e-07,
|
|
"loss": 0.5916,
|
|
"step": 1037
|
|
},
|
|
{
|
|
"epoch": 3.44,
|
|
"eval_loss": 0.7818962931632996,
|
|
"eval_runtime": 17.6137,
|
|
"eval_samples_per_second": 129.558,
|
|
"eval_steps_per_second": 5.45,
|
|
"step": 1037
|
|
},
|
|
{
|
|
"epoch": 3.44,
|
|
"learning_rate": 9.07608695652174e-07,
|
|
"loss": 0.5816,
|
|
"step": 1038
|
|
},
|
|
{
|
|
"epoch": 3.45,
|
|
"learning_rate": 9.021739130434782e-07,
|
|
"loss": 0.5946,
|
|
"step": 1039
|
|
},
|
|
{
|
|
"epoch": 3.45,
|
|
"learning_rate": 8.967391304347826e-07,
|
|
"loss": 0.5826,
|
|
"step": 1040
|
|
},
|
|
{
|
|
"epoch": 3.45,
|
|
"learning_rate": 8.913043478260869e-07,
|
|
"loss": 0.5805,
|
|
"step": 1041
|
|
},
|
|
{
|
|
"epoch": 3.46,
|
|
"learning_rate": 8.858695652173913e-07,
|
|
"loss": 0.6098,
|
|
"step": 1042
|
|
},
|
|
{
|
|
"epoch": 3.46,
|
|
"learning_rate": 8.804347826086957e-07,
|
|
"loss": 0.5869,
|
|
"step": 1043
|
|
},
|
|
{
|
|
"epoch": 3.46,
|
|
"learning_rate": 8.750000000000001e-07,
|
|
"loss": 0.573,
|
|
"step": 1044
|
|
},
|
|
{
|
|
"epoch": 3.47,
|
|
"learning_rate": 8.695652173913044e-07,
|
|
"loss": 0.5862,
|
|
"step": 1045
|
|
},
|
|
{
|
|
"epoch": 3.47,
|
|
"learning_rate": 8.641304347826088e-07,
|
|
"loss": 0.5944,
|
|
"step": 1046
|
|
},
|
|
{
|
|
"epoch": 3.47,
|
|
"learning_rate": 8.586956521739131e-07,
|
|
"loss": 0.5824,
|
|
"step": 1047
|
|
},
|
|
{
|
|
"epoch": 3.48,
|
|
"learning_rate": 8.532608695652173e-07,
|
|
"loss": 0.5997,
|
|
"step": 1048
|
|
},
|
|
{
|
|
"epoch": 3.48,
|
|
"learning_rate": 8.478260869565217e-07,
|
|
"loss": 0.5699,
|
|
"step": 1049
|
|
},
|
|
{
|
|
"epoch": 3.48,
|
|
"learning_rate": 8.423913043478262e-07,
|
|
"loss": 0.5964,
|
|
"step": 1050
|
|
},
|
|
{
|
|
"epoch": 3.49,
|
|
"learning_rate": 8.369565217391305e-07,
|
|
"loss": 0.5931,
|
|
"step": 1051
|
|
},
|
|
{
|
|
"epoch": 3.49,
|
|
"learning_rate": 8.315217391304348e-07,
|
|
"loss": 0.5951,
|
|
"step": 1052
|
|
},
|
|
{
|
|
"epoch": 3.49,
|
|
"learning_rate": 8.260869565217392e-07,
|
|
"loss": 0.5769,
|
|
"step": 1053
|
|
},
|
|
{
|
|
"epoch": 3.5,
|
|
"learning_rate": 8.206521739130435e-07,
|
|
"loss": 0.5687,
|
|
"step": 1054
|
|
},
|
|
{
|
|
"epoch": 3.5,
|
|
"learning_rate": 8.152173913043478e-07,
|
|
"loss": 0.5932,
|
|
"step": 1055
|
|
},
|
|
{
|
|
"epoch": 3.5,
|
|
"learning_rate": 8.097826086956521e-07,
|
|
"loss": 0.6009,
|
|
"step": 1056
|
|
},
|
|
{
|
|
"epoch": 3.51,
|
|
"learning_rate": 8.043478260869566e-07,
|
|
"loss": 0.5691,
|
|
"step": 1057
|
|
},
|
|
{
|
|
"epoch": 3.51,
|
|
"learning_rate": 7.989130434782609e-07,
|
|
"loss": 0.593,
|
|
"step": 1058
|
|
},
|
|
{
|
|
"epoch": 3.51,
|
|
"learning_rate": 7.934782608695652e-07,
|
|
"loss": 0.59,
|
|
"step": 1059
|
|
},
|
|
{
|
|
"epoch": 3.52,
|
|
"learning_rate": 7.880434782608696e-07,
|
|
"loss": 0.5868,
|
|
"step": 1060
|
|
},
|
|
{
|
|
"epoch": 3.52,
|
|
"learning_rate": 7.826086956521739e-07,
|
|
"loss": 0.588,
|
|
"step": 1061
|
|
},
|
|
{
|
|
"epoch": 3.52,
|
|
"learning_rate": 7.771739130434783e-07,
|
|
"loss": 0.5839,
|
|
"step": 1062
|
|
},
|
|
{
|
|
"epoch": 3.53,
|
|
"learning_rate": 7.717391304347826e-07,
|
|
"loss": 0.5876,
|
|
"step": 1063
|
|
},
|
|
{
|
|
"epoch": 3.53,
|
|
"learning_rate": 7.663043478260871e-07,
|
|
"loss": 0.5799,
|
|
"step": 1064
|
|
},
|
|
{
|
|
"epoch": 3.53,
|
|
"learning_rate": 7.608695652173913e-07,
|
|
"loss": 0.5802,
|
|
"step": 1065
|
|
},
|
|
{
|
|
"epoch": 3.54,
|
|
"learning_rate": 7.554347826086957e-07,
|
|
"loss": 0.593,
|
|
"step": 1066
|
|
},
|
|
{
|
|
"epoch": 3.54,
|
|
"learning_rate": 7.5e-07,
|
|
"loss": 0.6039,
|
|
"step": 1067
|
|
},
|
|
{
|
|
"epoch": 3.54,
|
|
"learning_rate": 7.445652173913043e-07,
|
|
"loss": 0.5731,
|
|
"step": 1068
|
|
},
|
|
{
|
|
"epoch": 3.55,
|
|
"learning_rate": 7.391304347826088e-07,
|
|
"loss": 0.5941,
|
|
"step": 1069
|
|
},
|
|
{
|
|
"epoch": 3.55,
|
|
"learning_rate": 7.336956521739131e-07,
|
|
"loss": 0.6007,
|
|
"step": 1070
|
|
},
|
|
{
|
|
"epoch": 3.55,
|
|
"learning_rate": 7.282608695652174e-07,
|
|
"loss": 0.5724,
|
|
"step": 1071
|
|
},
|
|
{
|
|
"epoch": 3.56,
|
|
"learning_rate": 7.228260869565218e-07,
|
|
"loss": 0.5829,
|
|
"step": 1072
|
|
},
|
|
{
|
|
"epoch": 3.56,
|
|
"learning_rate": 7.173913043478262e-07,
|
|
"loss": 0.5802,
|
|
"step": 1073
|
|
},
|
|
{
|
|
"epoch": 3.56,
|
|
"learning_rate": 7.119565217391304e-07,
|
|
"loss": 0.591,
|
|
"step": 1074
|
|
},
|
|
{
|
|
"epoch": 3.57,
|
|
"learning_rate": 7.065217391304348e-07,
|
|
"loss": 0.5748,
|
|
"step": 1075
|
|
},
|
|
{
|
|
"epoch": 3.57,
|
|
"learning_rate": 7.010869565217392e-07,
|
|
"loss": 0.5877,
|
|
"step": 1076
|
|
},
|
|
{
|
|
"epoch": 3.57,
|
|
"learning_rate": 6.956521739130435e-07,
|
|
"loss": 0.6029,
|
|
"step": 1077
|
|
},
|
|
{
|
|
"epoch": 3.58,
|
|
"learning_rate": 6.902173913043478e-07,
|
|
"loss": 0.5827,
|
|
"step": 1078
|
|
},
|
|
{
|
|
"epoch": 3.58,
|
|
"learning_rate": 6.847826086956522e-07,
|
|
"loss": 0.5958,
|
|
"step": 1079
|
|
},
|
|
{
|
|
"epoch": 3.58,
|
|
"learning_rate": 6.793478260869566e-07,
|
|
"loss": 0.569,
|
|
"step": 1080
|
|
},
|
|
{
|
|
"epoch": 3.59,
|
|
"learning_rate": 6.739130434782609e-07,
|
|
"loss": 0.5899,
|
|
"step": 1081
|
|
},
|
|
{
|
|
"epoch": 3.59,
|
|
"learning_rate": 6.684782608695652e-07,
|
|
"loss": 0.5876,
|
|
"step": 1082
|
|
},
|
|
{
|
|
"epoch": 3.59,
|
|
"learning_rate": 6.630434782608696e-07,
|
|
"loss": 0.5833,
|
|
"step": 1083
|
|
},
|
|
{
|
|
"epoch": 3.6,
|
|
"learning_rate": 6.576086956521739e-07,
|
|
"loss": 0.5949,
|
|
"step": 1084
|
|
},
|
|
{
|
|
"epoch": 3.6,
|
|
"learning_rate": 6.521739130434783e-07,
|
|
"loss": 0.5917,
|
|
"step": 1085
|
|
},
|
|
{
|
|
"epoch": 3.6,
|
|
"learning_rate": 6.467391304347827e-07,
|
|
"loss": 0.5877,
|
|
"step": 1086
|
|
},
|
|
{
|
|
"epoch": 3.61,
|
|
"learning_rate": 6.41304347826087e-07,
|
|
"loss": 0.5826,
|
|
"step": 1087
|
|
},
|
|
{
|
|
"epoch": 3.61,
|
|
"learning_rate": 6.358695652173913e-07,
|
|
"loss": 0.5922,
|
|
"step": 1088
|
|
},
|
|
{
|
|
"epoch": 3.61,
|
|
"learning_rate": 6.304347826086957e-07,
|
|
"loss": 0.5796,
|
|
"step": 1089
|
|
},
|
|
{
|
|
"epoch": 3.62,
|
|
"learning_rate": 6.25e-07,
|
|
"loss": 0.5744,
|
|
"step": 1090
|
|
},
|
|
{
|
|
"epoch": 3.62,
|
|
"learning_rate": 6.195652173913043e-07,
|
|
"loss": 0.5875,
|
|
"step": 1091
|
|
},
|
|
{
|
|
"epoch": 3.62,
|
|
"learning_rate": 6.141304347826087e-07,
|
|
"loss": 0.5811,
|
|
"step": 1092
|
|
},
|
|
{
|
|
"epoch": 3.63,
|
|
"learning_rate": 6.086956521739131e-07,
|
|
"loss": 0.5816,
|
|
"step": 1093
|
|
},
|
|
{
|
|
"epoch": 3.63,
|
|
"learning_rate": 6.032608695652174e-07,
|
|
"loss": 0.5956,
|
|
"step": 1094
|
|
},
|
|
{
|
|
"epoch": 3.63,
|
|
"learning_rate": 5.978260869565218e-07,
|
|
"loss": 0.5916,
|
|
"step": 1095
|
|
},
|
|
{
|
|
"epoch": 3.64,
|
|
"learning_rate": 5.923913043478261e-07,
|
|
"loss": 0.5795,
|
|
"step": 1096
|
|
},
|
|
{
|
|
"epoch": 3.64,
|
|
"learning_rate": 5.869565217391305e-07,
|
|
"loss": 0.5878,
|
|
"step": 1097
|
|
},
|
|
{
|
|
"epoch": 3.64,
|
|
"learning_rate": 5.815217391304348e-07,
|
|
"loss": 0.6141,
|
|
"step": 1098
|
|
},
|
|
{
|
|
"epoch": 3.64,
|
|
"eval_loss": 0.7818270325660706,
|
|
"eval_runtime": 17.6197,
|
|
"eval_samples_per_second": 129.514,
|
|
"eval_steps_per_second": 5.448,
|
|
"step": 1098
|
|
},
|
|
{
|
|
"epoch": 3.65,
|
|
"learning_rate": 5.760869565217391e-07,
|
|
"loss": 0.5944,
|
|
"step": 1099
|
|
},
|
|
{
|
|
"epoch": 3.65,
|
|
"learning_rate": 5.706521739130435e-07,
|
|
"loss": 0.5901,
|
|
"step": 1100
|
|
},
|
|
{
|
|
"epoch": 3.65,
|
|
"learning_rate": 5.652173913043478e-07,
|
|
"loss": 0.5932,
|
|
"step": 1101
|
|
},
|
|
{
|
|
"epoch": 3.66,
|
|
"learning_rate": 5.597826086956522e-07,
|
|
"loss": 0.5929,
|
|
"step": 1102
|
|
},
|
|
{
|
|
"epoch": 3.66,
|
|
"learning_rate": 5.543478260869565e-07,
|
|
"loss": 0.5674,
|
|
"step": 1103
|
|
},
|
|
{
|
|
"epoch": 3.66,
|
|
"learning_rate": 5.489130434782609e-07,
|
|
"loss": 0.5765,
|
|
"step": 1104
|
|
},
|
|
{
|
|
"epoch": 3.67,
|
|
"learning_rate": 5.434782608695653e-07,
|
|
"loss": 0.5847,
|
|
"step": 1105
|
|
},
|
|
{
|
|
"epoch": 3.67,
|
|
"learning_rate": 5.380434782608696e-07,
|
|
"loss": 0.6048,
|
|
"step": 1106
|
|
},
|
|
{
|
|
"epoch": 3.67,
|
|
"learning_rate": 5.32608695652174e-07,
|
|
"loss": 0.5921,
|
|
"step": 1107
|
|
},
|
|
{
|
|
"epoch": 3.67,
|
|
"learning_rate": 5.271739130434782e-07,
|
|
"loss": 0.5985,
|
|
"step": 1108
|
|
},
|
|
{
|
|
"epoch": 3.68,
|
|
"learning_rate": 5.217391304347826e-07,
|
|
"loss": 0.5727,
|
|
"step": 1109
|
|
},
|
|
{
|
|
"epoch": 3.68,
|
|
"learning_rate": 5.163043478260869e-07,
|
|
"loss": 0.5807,
|
|
"step": 1110
|
|
},
|
|
{
|
|
"epoch": 3.68,
|
|
"learning_rate": 5.108695652173913e-07,
|
|
"loss": 0.5804,
|
|
"step": 1111
|
|
},
|
|
{
|
|
"epoch": 3.69,
|
|
"learning_rate": 5.054347826086957e-07,
|
|
"loss": 0.5789,
|
|
"step": 1112
|
|
},
|
|
{
|
|
"epoch": 3.69,
|
|
"learning_rate": 5e-07,
|
|
"loss": 0.5782,
|
|
"step": 1113
|
|
},
|
|
{
|
|
"epoch": 3.69,
|
|
"learning_rate": 4.945652173913044e-07,
|
|
"loss": 0.5757,
|
|
"step": 1114
|
|
},
|
|
{
|
|
"epoch": 3.7,
|
|
"learning_rate": 4.891304347826088e-07,
|
|
"loss": 0.6007,
|
|
"step": 1115
|
|
},
|
|
{
|
|
"epoch": 3.7,
|
|
"learning_rate": 4.83695652173913e-07,
|
|
"loss": 0.5844,
|
|
"step": 1116
|
|
},
|
|
{
|
|
"epoch": 3.7,
|
|
"learning_rate": 4.782608695652173e-07,
|
|
"loss": 0.5973,
|
|
"step": 1117
|
|
},
|
|
{
|
|
"epoch": 3.71,
|
|
"learning_rate": 4.7282608695652177e-07,
|
|
"loss": 0.5796,
|
|
"step": 1118
|
|
},
|
|
{
|
|
"epoch": 3.71,
|
|
"learning_rate": 4.673913043478261e-07,
|
|
"loss": 0.5806,
|
|
"step": 1119
|
|
},
|
|
{
|
|
"epoch": 3.71,
|
|
"learning_rate": 4.619565217391304e-07,
|
|
"loss": 0.5921,
|
|
"step": 1120
|
|
},
|
|
{
|
|
"epoch": 3.72,
|
|
"learning_rate": 4.5652173913043484e-07,
|
|
"loss": 0.5825,
|
|
"step": 1121
|
|
},
|
|
{
|
|
"epoch": 3.72,
|
|
"learning_rate": 4.510869565217391e-07,
|
|
"loss": 0.5784,
|
|
"step": 1122
|
|
},
|
|
{
|
|
"epoch": 3.72,
|
|
"learning_rate": 4.4565217391304346e-07,
|
|
"loss": 0.5956,
|
|
"step": 1123
|
|
},
|
|
{
|
|
"epoch": 3.73,
|
|
"learning_rate": 4.4021739130434785e-07,
|
|
"loss": 0.585,
|
|
"step": 1124
|
|
},
|
|
{
|
|
"epoch": 3.73,
|
|
"learning_rate": 4.347826086956522e-07,
|
|
"loss": 0.5804,
|
|
"step": 1125
|
|
},
|
|
{
|
|
"epoch": 3.73,
|
|
"learning_rate": 4.2934782608695653e-07,
|
|
"loss": 0.6056,
|
|
"step": 1126
|
|
},
|
|
{
|
|
"epoch": 3.74,
|
|
"learning_rate": 4.2391304347826086e-07,
|
|
"loss": 0.5867,
|
|
"step": 1127
|
|
},
|
|
{
|
|
"epoch": 3.74,
|
|
"learning_rate": 4.1847826086956525e-07,
|
|
"loss": 0.5887,
|
|
"step": 1128
|
|
},
|
|
{
|
|
"epoch": 3.74,
|
|
"learning_rate": 4.130434782608696e-07,
|
|
"loss": 0.565,
|
|
"step": 1129
|
|
},
|
|
{
|
|
"epoch": 3.75,
|
|
"learning_rate": 4.076086956521739e-07,
|
|
"loss": 0.567,
|
|
"step": 1130
|
|
},
|
|
{
|
|
"epoch": 3.75,
|
|
"learning_rate": 4.021739130434783e-07,
|
|
"loss": 0.5955,
|
|
"step": 1131
|
|
},
|
|
{
|
|
"epoch": 3.75,
|
|
"learning_rate": 3.967391304347826e-07,
|
|
"loss": 0.5934,
|
|
"step": 1132
|
|
},
|
|
{
|
|
"epoch": 3.76,
|
|
"learning_rate": 3.9130434782608694e-07,
|
|
"loss": 0.5733,
|
|
"step": 1133
|
|
},
|
|
{
|
|
"epoch": 3.76,
|
|
"learning_rate": 3.858695652173913e-07,
|
|
"loss": 0.6031,
|
|
"step": 1134
|
|
},
|
|
{
|
|
"epoch": 3.76,
|
|
"learning_rate": 3.8043478260869567e-07,
|
|
"loss": 0.5827,
|
|
"step": 1135
|
|
},
|
|
{
|
|
"epoch": 3.77,
|
|
"learning_rate": 3.75e-07,
|
|
"loss": 0.5879,
|
|
"step": 1136
|
|
},
|
|
{
|
|
"epoch": 3.77,
|
|
"learning_rate": 3.695652173913044e-07,
|
|
"loss": 0.5773,
|
|
"step": 1137
|
|
},
|
|
{
|
|
"epoch": 3.77,
|
|
"learning_rate": 3.641304347826087e-07,
|
|
"loss": 0.5823,
|
|
"step": 1138
|
|
},
|
|
{
|
|
"epoch": 3.78,
|
|
"learning_rate": 3.586956521739131e-07,
|
|
"loss": 0.5853,
|
|
"step": 1139
|
|
},
|
|
{
|
|
"epoch": 3.78,
|
|
"learning_rate": 3.532608695652174e-07,
|
|
"loss": 0.5871,
|
|
"step": 1140
|
|
},
|
|
{
|
|
"epoch": 3.78,
|
|
"learning_rate": 3.4782608695652175e-07,
|
|
"loss": 0.595,
|
|
"step": 1141
|
|
},
|
|
{
|
|
"epoch": 3.79,
|
|
"learning_rate": 3.423913043478261e-07,
|
|
"loss": 0.5915,
|
|
"step": 1142
|
|
},
|
|
{
|
|
"epoch": 3.79,
|
|
"learning_rate": 3.369565217391304e-07,
|
|
"loss": 0.581,
|
|
"step": 1143
|
|
},
|
|
{
|
|
"epoch": 3.79,
|
|
"learning_rate": 3.315217391304348e-07,
|
|
"loss": 0.5883,
|
|
"step": 1144
|
|
},
|
|
{
|
|
"epoch": 3.8,
|
|
"learning_rate": 3.2608695652173915e-07,
|
|
"loss": 0.573,
|
|
"step": 1145
|
|
},
|
|
{
|
|
"epoch": 3.8,
|
|
"learning_rate": 3.206521739130435e-07,
|
|
"loss": 0.5737,
|
|
"step": 1146
|
|
},
|
|
{
|
|
"epoch": 3.8,
|
|
"learning_rate": 3.1521739130434783e-07,
|
|
"loss": 0.5638,
|
|
"step": 1147
|
|
},
|
|
{
|
|
"epoch": 3.81,
|
|
"learning_rate": 3.0978260869565217e-07,
|
|
"loss": 0.5953,
|
|
"step": 1148
|
|
},
|
|
{
|
|
"epoch": 3.81,
|
|
"learning_rate": 3.0434782608695656e-07,
|
|
"loss": 0.5984,
|
|
"step": 1149
|
|
},
|
|
{
|
|
"epoch": 3.81,
|
|
"learning_rate": 2.989130434782609e-07,
|
|
"loss": 0.5913,
|
|
"step": 1150
|
|
},
|
|
{
|
|
"epoch": 3.82,
|
|
"learning_rate": 2.9347826086956523e-07,
|
|
"loss": 0.5957,
|
|
"step": 1151
|
|
},
|
|
{
|
|
"epoch": 3.82,
|
|
"learning_rate": 2.8804347826086957e-07,
|
|
"loss": 0.5824,
|
|
"step": 1152
|
|
},
|
|
{
|
|
"epoch": 3.82,
|
|
"learning_rate": 2.826086956521739e-07,
|
|
"loss": 0.5823,
|
|
"step": 1153
|
|
},
|
|
{
|
|
"epoch": 3.83,
|
|
"learning_rate": 2.7717391304347825e-07,
|
|
"loss": 0.5796,
|
|
"step": 1154
|
|
},
|
|
{
|
|
"epoch": 3.83,
|
|
"learning_rate": 2.7173913043478264e-07,
|
|
"loss": 0.5915,
|
|
"step": 1155
|
|
},
|
|
{
|
|
"epoch": 3.83,
|
|
"learning_rate": 2.66304347826087e-07,
|
|
"loss": 0.6044,
|
|
"step": 1156
|
|
},
|
|
{
|
|
"epoch": 3.84,
|
|
"learning_rate": 2.608695652173913e-07,
|
|
"loss": 0.5918,
|
|
"step": 1157
|
|
},
|
|
{
|
|
"epoch": 3.84,
|
|
"learning_rate": 2.5543478260869565e-07,
|
|
"loss": 0.5903,
|
|
"step": 1158
|
|
},
|
|
{
|
|
"epoch": 3.84,
|
|
"learning_rate": 2.5e-07,
|
|
"loss": 0.595,
|
|
"step": 1159
|
|
},
|
|
{
|
|
"epoch": 3.84,
|
|
"eval_loss": 0.7819833755493164,
|
|
"eval_runtime": 17.6271,
|
|
"eval_samples_per_second": 129.46,
|
|
"eval_steps_per_second": 5.446,
|
|
"step": 1159
|
|
},
|
|
{
|
|
"epoch": 3.85,
|
|
"learning_rate": 2.445652173913044e-07,
|
|
"loss": 0.5926,
|
|
"step": 1160
|
|
},
|
|
{
|
|
"epoch": 3.85,
|
|
"learning_rate": 2.3913043478260866e-07,
|
|
"loss": 0.5802,
|
|
"step": 1161
|
|
},
|
|
{
|
|
"epoch": 3.85,
|
|
"learning_rate": 2.3369565217391305e-07,
|
|
"loss": 0.5726,
|
|
"step": 1162
|
|
},
|
|
{
|
|
"epoch": 3.86,
|
|
"learning_rate": 2.2826086956521742e-07,
|
|
"loss": 0.592,
|
|
"step": 1163
|
|
},
|
|
{
|
|
"epoch": 3.86,
|
|
"learning_rate": 2.2282608695652173e-07,
|
|
"loss": 0.6058,
|
|
"step": 1164
|
|
},
|
|
{
|
|
"epoch": 3.86,
|
|
"learning_rate": 2.173913043478261e-07,
|
|
"loss": 0.5843,
|
|
"step": 1165
|
|
},
|
|
{
|
|
"epoch": 3.87,
|
|
"learning_rate": 2.1195652173913043e-07,
|
|
"loss": 0.5848,
|
|
"step": 1166
|
|
},
|
|
{
|
|
"epoch": 3.87,
|
|
"learning_rate": 2.065217391304348e-07,
|
|
"loss": 0.603,
|
|
"step": 1167
|
|
},
|
|
{
|
|
"epoch": 3.87,
|
|
"learning_rate": 2.0108695652173916e-07,
|
|
"loss": 0.5937,
|
|
"step": 1168
|
|
},
|
|
{
|
|
"epoch": 3.88,
|
|
"learning_rate": 1.9565217391304347e-07,
|
|
"loss": 0.573,
|
|
"step": 1169
|
|
},
|
|
{
|
|
"epoch": 3.88,
|
|
"learning_rate": 1.9021739130434784e-07,
|
|
"loss": 0.5811,
|
|
"step": 1170
|
|
},
|
|
{
|
|
"epoch": 3.88,
|
|
"learning_rate": 1.847826086956522e-07,
|
|
"loss": 0.5945,
|
|
"step": 1171
|
|
},
|
|
{
|
|
"epoch": 3.89,
|
|
"learning_rate": 1.7934782608695654e-07,
|
|
"loss": 0.5802,
|
|
"step": 1172
|
|
},
|
|
{
|
|
"epoch": 3.89,
|
|
"learning_rate": 1.7391304347826088e-07,
|
|
"loss": 0.5944,
|
|
"step": 1173
|
|
},
|
|
{
|
|
"epoch": 3.89,
|
|
"learning_rate": 1.684782608695652e-07,
|
|
"loss": 0.5901,
|
|
"step": 1174
|
|
},
|
|
{
|
|
"epoch": 3.9,
|
|
"learning_rate": 1.6304347826086958e-07,
|
|
"loss": 0.5786,
|
|
"step": 1175
|
|
},
|
|
{
|
|
"epoch": 3.9,
|
|
"learning_rate": 1.5760869565217392e-07,
|
|
"loss": 0.5843,
|
|
"step": 1176
|
|
},
|
|
{
|
|
"epoch": 3.9,
|
|
"learning_rate": 1.5217391304347828e-07,
|
|
"loss": 0.5771,
|
|
"step": 1177
|
|
},
|
|
{
|
|
"epoch": 3.91,
|
|
"learning_rate": 1.4673913043478262e-07,
|
|
"loss": 0.5921,
|
|
"step": 1178
|
|
},
|
|
{
|
|
"epoch": 3.91,
|
|
"learning_rate": 1.4130434782608695e-07,
|
|
"loss": 0.577,
|
|
"step": 1179
|
|
},
|
|
{
|
|
"epoch": 3.91,
|
|
"learning_rate": 1.3586956521739132e-07,
|
|
"loss": 0.5844,
|
|
"step": 1180
|
|
},
|
|
{
|
|
"epoch": 3.92,
|
|
"learning_rate": 1.3043478260869566e-07,
|
|
"loss": 0.5803,
|
|
"step": 1181
|
|
},
|
|
{
|
|
"epoch": 3.92,
|
|
"learning_rate": 1.25e-07,
|
|
"loss": 0.5698,
|
|
"step": 1182
|
|
},
|
|
{
|
|
"epoch": 3.92,
|
|
"learning_rate": 1.1956521739130433e-07,
|
|
"loss": 0.605,
|
|
"step": 1183
|
|
},
|
|
{
|
|
"epoch": 3.93,
|
|
"learning_rate": 1.1413043478260871e-07,
|
|
"loss": 0.5873,
|
|
"step": 1184
|
|
},
|
|
{
|
|
"epoch": 3.93,
|
|
"learning_rate": 1.0869565217391305e-07,
|
|
"loss": 0.5849,
|
|
"step": 1185
|
|
},
|
|
{
|
|
"epoch": 3.93,
|
|
"learning_rate": 1.032608695652174e-07,
|
|
"loss": 0.5926,
|
|
"step": 1186
|
|
},
|
|
{
|
|
"epoch": 3.94,
|
|
"learning_rate": 9.782608695652174e-08,
|
|
"loss": 0.5804,
|
|
"step": 1187
|
|
},
|
|
{
|
|
"epoch": 3.94,
|
|
"learning_rate": 9.23913043478261e-08,
|
|
"loss": 0.5786,
|
|
"step": 1188
|
|
},
|
|
{
|
|
"epoch": 3.94,
|
|
"learning_rate": 8.695652173913044e-08,
|
|
"loss": 0.582,
|
|
"step": 1189
|
|
},
|
|
{
|
|
"epoch": 3.95,
|
|
"learning_rate": 8.152173913043479e-08,
|
|
"loss": 0.578,
|
|
"step": 1190
|
|
},
|
|
{
|
|
"epoch": 3.95,
|
|
"learning_rate": 7.608695652173914e-08,
|
|
"loss": 0.583,
|
|
"step": 1191
|
|
},
|
|
{
|
|
"epoch": 3.95,
|
|
"learning_rate": 7.065217391304348e-08,
|
|
"loss": 0.5603,
|
|
"step": 1192
|
|
},
|
|
{
|
|
"epoch": 3.96,
|
|
"learning_rate": 6.521739130434783e-08,
|
|
"loss": 0.5908,
|
|
"step": 1193
|
|
},
|
|
{
|
|
"epoch": 3.96,
|
|
"learning_rate": 5.978260869565217e-08,
|
|
"loss": 0.5871,
|
|
"step": 1194
|
|
},
|
|
{
|
|
"epoch": 3.96,
|
|
"learning_rate": 5.4347826086956524e-08,
|
|
"loss": 0.5958,
|
|
"step": 1195
|
|
},
|
|
{
|
|
"epoch": 3.97,
|
|
"learning_rate": 4.891304347826087e-08,
|
|
"loss": 0.588,
|
|
"step": 1196
|
|
},
|
|
{
|
|
"epoch": 3.97,
|
|
"learning_rate": 4.347826086956522e-08,
|
|
"loss": 0.5838,
|
|
"step": 1197
|
|
},
|
|
{
|
|
"epoch": 3.97,
|
|
"learning_rate": 3.804347826086957e-08,
|
|
"loss": 0.5835,
|
|
"step": 1198
|
|
},
|
|
{
|
|
"epoch": 3.98,
|
|
"learning_rate": 3.2608695652173914e-08,
|
|
"loss": 0.5801,
|
|
"step": 1199
|
|
},
|
|
{
|
|
"epoch": 3.98,
|
|
"learning_rate": 2.7173913043478262e-08,
|
|
"loss": 0.5959,
|
|
"step": 1200
|
|
},
|
|
{
|
|
"epoch": 3.98,
|
|
"learning_rate": 2.173913043478261e-08,
|
|
"loss": 0.5649,
|
|
"step": 1201
|
|
},
|
|
{
|
|
"epoch": 3.99,
|
|
"learning_rate": 1.6304347826086957e-08,
|
|
"loss": 0.5745,
|
|
"step": 1202
|
|
},
|
|
{
|
|
"epoch": 3.99,
|
|
"learning_rate": 1.0869565217391305e-08,
|
|
"loss": 0.5876,
|
|
"step": 1203
|
|
},
|
|
{
|
|
"epoch": 3.99,
|
|
"learning_rate": 5.434782608695652e-09,
|
|
"loss": 0.5869,
|
|
"step": 1204
|
|
}
|
|
],
|
|
"logging_steps": 1,
|
|
"max_steps": 1204,
|
|
"num_train_epochs": 4,
|
|
"save_steps": 500,
|
|
"total_flos": 4.039702036808624e+19,
|
|
"trial_name": null,
|
|
"trial_params": null
|
|
}
|