4383 lines
106 KiB
JSON
4383 lines
106 KiB
JSON
{
|
|
"best_metric": null,
|
|
"best_model_checkpoint": null,
|
|
"epoch": 1.993968636911942,
|
|
"eval_steps": 500,
|
|
"global_step": 620,
|
|
"is_hyper_param_search": false,
|
|
"is_local_process_zero": true,
|
|
"is_world_process_zero": true,
|
|
"log_history": [
|
|
{
|
|
"epoch": 0.003216726980297547,
|
|
"grad_norm": 81.91886138916016,
|
|
"learning_rate": 1.6129032258064518e-07,
|
|
"loss": 0.3558,
|
|
"step": 1
|
|
},
|
|
{
|
|
"epoch": 0.006433453960595094,
|
|
"grad_norm": 88.99775695800781,
|
|
"learning_rate": 3.2258064516129035e-07,
|
|
"loss": 0.4071,
|
|
"step": 2
|
|
},
|
|
{
|
|
"epoch": 0.009650180940892641,
|
|
"grad_norm": 79.32195281982422,
|
|
"learning_rate": 4.838709677419355e-07,
|
|
"loss": 0.3576,
|
|
"step": 3
|
|
},
|
|
{
|
|
"epoch": 0.012866907921190189,
|
|
"grad_norm": 81.10282897949219,
|
|
"learning_rate": 6.451612903225807e-07,
|
|
"loss": 0.3627,
|
|
"step": 4
|
|
},
|
|
{
|
|
"epoch": 0.016083634901487735,
|
|
"grad_norm": 75.47724914550781,
|
|
"learning_rate": 8.064516129032258e-07,
|
|
"loss": 0.3581,
|
|
"step": 5
|
|
},
|
|
{
|
|
"epoch": 0.019300361881785282,
|
|
"grad_norm": 73.94715118408203,
|
|
"learning_rate": 9.67741935483871e-07,
|
|
"loss": 0.3392,
|
|
"step": 6
|
|
},
|
|
{
|
|
"epoch": 0.02251708886208283,
|
|
"grad_norm": 54.852840423583984,
|
|
"learning_rate": 1.1290322580645162e-06,
|
|
"loss": 0.2786,
|
|
"step": 7
|
|
},
|
|
{
|
|
"epoch": 0.025733815842380377,
|
|
"grad_norm": 51.52576446533203,
|
|
"learning_rate": 1.2903225806451614e-06,
|
|
"loss": 0.27,
|
|
"step": 8
|
|
},
|
|
{
|
|
"epoch": 0.028950542822677925,
|
|
"grad_norm": 17.170745849609375,
|
|
"learning_rate": 1.4516129032258066e-06,
|
|
"loss": 0.1758,
|
|
"step": 9
|
|
},
|
|
{
|
|
"epoch": 0.03216726980297547,
|
|
"grad_norm": 10.435275077819824,
|
|
"learning_rate": 1.6129032258064516e-06,
|
|
"loss": 0.1723,
|
|
"step": 10
|
|
},
|
|
{
|
|
"epoch": 0.03538399678327302,
|
|
"grad_norm": 8.09239387512207,
|
|
"learning_rate": 1.774193548387097e-06,
|
|
"loss": 0.1565,
|
|
"step": 11
|
|
},
|
|
{
|
|
"epoch": 0.038600723763570564,
|
|
"grad_norm": 41.39366149902344,
|
|
"learning_rate": 1.935483870967742e-06,
|
|
"loss": 0.1746,
|
|
"step": 12
|
|
},
|
|
{
|
|
"epoch": 0.04181745074386811,
|
|
"grad_norm": 20.37698745727539,
|
|
"learning_rate": 2.096774193548387e-06,
|
|
"loss": 0.1709,
|
|
"step": 13
|
|
},
|
|
{
|
|
"epoch": 0.04503417772416566,
|
|
"grad_norm": 10.368378639221191,
|
|
"learning_rate": 2.2580645161290324e-06,
|
|
"loss": 0.1567,
|
|
"step": 14
|
|
},
|
|
{
|
|
"epoch": 0.04825090470446321,
|
|
"grad_norm": 14.423890113830566,
|
|
"learning_rate": 2.4193548387096776e-06,
|
|
"loss": 0.1344,
|
|
"step": 15
|
|
},
|
|
{
|
|
"epoch": 0.051467631684760755,
|
|
"grad_norm": 13.449414253234863,
|
|
"learning_rate": 2.580645161290323e-06,
|
|
"loss": 0.1306,
|
|
"step": 16
|
|
},
|
|
{
|
|
"epoch": 0.0546843586650583,
|
|
"grad_norm": 5.699060916900635,
|
|
"learning_rate": 2.7419354838709676e-06,
|
|
"loss": 0.1368,
|
|
"step": 17
|
|
},
|
|
{
|
|
"epoch": 0.05790108564535585,
|
|
"grad_norm": 38.163658142089844,
|
|
"learning_rate": 2.903225806451613e-06,
|
|
"loss": 0.1767,
|
|
"step": 18
|
|
},
|
|
{
|
|
"epoch": 0.0611178126256534,
|
|
"grad_norm": 43.165924072265625,
|
|
"learning_rate": 3.0645161290322584e-06,
|
|
"loss": 0.2103,
|
|
"step": 19
|
|
},
|
|
{
|
|
"epoch": 0.06433453960595094,
|
|
"grad_norm": 34.02187728881836,
|
|
"learning_rate": 3.225806451612903e-06,
|
|
"loss": 0.1798,
|
|
"step": 20
|
|
},
|
|
{
|
|
"epoch": 0.06755126658624849,
|
|
"grad_norm": 15.196747779846191,
|
|
"learning_rate": 3.3870967741935484e-06,
|
|
"loss": 0.1326,
|
|
"step": 21
|
|
},
|
|
{
|
|
"epoch": 0.07076799356654603,
|
|
"grad_norm": 12.960094451904297,
|
|
"learning_rate": 3.548387096774194e-06,
|
|
"loss": 0.132,
|
|
"step": 22
|
|
},
|
|
{
|
|
"epoch": 0.07398472054684359,
|
|
"grad_norm": 18.067081451416016,
|
|
"learning_rate": 3.7096774193548392e-06,
|
|
"loss": 0.1407,
|
|
"step": 23
|
|
},
|
|
{
|
|
"epoch": 0.07720144752714113,
|
|
"grad_norm": 11.63009262084961,
|
|
"learning_rate": 3.870967741935484e-06,
|
|
"loss": 0.1198,
|
|
"step": 24
|
|
},
|
|
{
|
|
"epoch": 0.08041817450743868,
|
|
"grad_norm": 3.6088430881500244,
|
|
"learning_rate": 4.032258064516129e-06,
|
|
"loss": 0.0861,
|
|
"step": 25
|
|
},
|
|
{
|
|
"epoch": 0.08363490148773622,
|
|
"grad_norm": 10.469372749328613,
|
|
"learning_rate": 4.193548387096774e-06,
|
|
"loss": 0.1043,
|
|
"step": 26
|
|
},
|
|
{
|
|
"epoch": 0.08685162846803378,
|
|
"grad_norm": 14.186717987060547,
|
|
"learning_rate": 4.35483870967742e-06,
|
|
"loss": 0.135,
|
|
"step": 27
|
|
},
|
|
{
|
|
"epoch": 0.09006835544833132,
|
|
"grad_norm": 7.1297454833984375,
|
|
"learning_rate": 4.516129032258065e-06,
|
|
"loss": 0.1171,
|
|
"step": 28
|
|
},
|
|
{
|
|
"epoch": 0.09328508242862887,
|
|
"grad_norm": 2.476020574569702,
|
|
"learning_rate": 4.67741935483871e-06,
|
|
"loss": 0.089,
|
|
"step": 29
|
|
},
|
|
{
|
|
"epoch": 0.09650180940892641,
|
|
"grad_norm": 3.7916111946105957,
|
|
"learning_rate": 4.838709677419355e-06,
|
|
"loss": 0.0881,
|
|
"step": 30
|
|
},
|
|
{
|
|
"epoch": 0.09971853638922397,
|
|
"grad_norm": 1.1642123460769653,
|
|
"learning_rate": 5e-06,
|
|
"loss": 0.0965,
|
|
"step": 31
|
|
},
|
|
{
|
|
"epoch": 0.10293526336952151,
|
|
"grad_norm": 3.4518535137176514,
|
|
"learning_rate": 5.161290322580646e-06,
|
|
"loss": 0.0849,
|
|
"step": 32
|
|
},
|
|
{
|
|
"epoch": 0.10615199034981906,
|
|
"grad_norm": 2.2884507179260254,
|
|
"learning_rate": 5.322580645161291e-06,
|
|
"loss": 0.0786,
|
|
"step": 33
|
|
},
|
|
{
|
|
"epoch": 0.1093687173301166,
|
|
"grad_norm": 1.3953402042388916,
|
|
"learning_rate": 5.483870967741935e-06,
|
|
"loss": 0.0675,
|
|
"step": 34
|
|
},
|
|
{
|
|
"epoch": 0.11258544431041416,
|
|
"grad_norm": 5.182158470153809,
|
|
"learning_rate": 5.645161290322582e-06,
|
|
"loss": 0.0977,
|
|
"step": 35
|
|
},
|
|
{
|
|
"epoch": 0.1158021712907117,
|
|
"grad_norm": 2.713275909423828,
|
|
"learning_rate": 5.806451612903226e-06,
|
|
"loss": 0.0648,
|
|
"step": 36
|
|
},
|
|
{
|
|
"epoch": 0.11901889827100925,
|
|
"grad_norm": 2.827744960784912,
|
|
"learning_rate": 5.967741935483872e-06,
|
|
"loss": 0.082,
|
|
"step": 37
|
|
},
|
|
{
|
|
"epoch": 0.1222356252513068,
|
|
"grad_norm": 4.1014723777771,
|
|
"learning_rate": 6.129032258064517e-06,
|
|
"loss": 0.0758,
|
|
"step": 38
|
|
},
|
|
{
|
|
"epoch": 0.12545235223160434,
|
|
"grad_norm": 1.5448578596115112,
|
|
"learning_rate": 6.290322580645162e-06,
|
|
"loss": 0.066,
|
|
"step": 39
|
|
},
|
|
{
|
|
"epoch": 0.12866907921190188,
|
|
"grad_norm": 3.3927855491638184,
|
|
"learning_rate": 6.451612903225806e-06,
|
|
"loss": 0.0709,
|
|
"step": 40
|
|
},
|
|
{
|
|
"epoch": 0.13188580619219945,
|
|
"grad_norm": 2.370821475982666,
|
|
"learning_rate": 6.612903225806452e-06,
|
|
"loss": 0.0852,
|
|
"step": 41
|
|
},
|
|
{
|
|
"epoch": 0.13510253317249699,
|
|
"grad_norm": 3.0705552101135254,
|
|
"learning_rate": 6.774193548387097e-06,
|
|
"loss": 0.0702,
|
|
"step": 42
|
|
},
|
|
{
|
|
"epoch": 0.13831926015279453,
|
|
"grad_norm": 4.215071678161621,
|
|
"learning_rate": 6.935483870967743e-06,
|
|
"loss": 0.0747,
|
|
"step": 43
|
|
},
|
|
{
|
|
"epoch": 0.14153598713309207,
|
|
"grad_norm": 2.137068748474121,
|
|
"learning_rate": 7.096774193548388e-06,
|
|
"loss": 0.0657,
|
|
"step": 44
|
|
},
|
|
{
|
|
"epoch": 0.14475271411338964,
|
|
"grad_norm": 4.441213130950928,
|
|
"learning_rate": 7.258064516129033e-06,
|
|
"loss": 0.0717,
|
|
"step": 45
|
|
},
|
|
{
|
|
"epoch": 0.14796944109368718,
|
|
"grad_norm": 2.896615743637085,
|
|
"learning_rate": 7.4193548387096784e-06,
|
|
"loss": 0.066,
|
|
"step": 46
|
|
},
|
|
{
|
|
"epoch": 0.15118616807398472,
|
|
"grad_norm": 1.594978928565979,
|
|
"learning_rate": 7.580645161290323e-06,
|
|
"loss": 0.0554,
|
|
"step": 47
|
|
},
|
|
{
|
|
"epoch": 0.15440289505428226,
|
|
"grad_norm": 3.3716137409210205,
|
|
"learning_rate": 7.741935483870968e-06,
|
|
"loss": 0.075,
|
|
"step": 48
|
|
},
|
|
{
|
|
"epoch": 0.15761962203457983,
|
|
"grad_norm": 2.1819281578063965,
|
|
"learning_rate": 7.903225806451613e-06,
|
|
"loss": 0.0583,
|
|
"step": 49
|
|
},
|
|
{
|
|
"epoch": 0.16083634901487737,
|
|
"grad_norm": 5.265512943267822,
|
|
"learning_rate": 8.064516129032258e-06,
|
|
"loss": 0.0832,
|
|
"step": 50
|
|
},
|
|
{
|
|
"epoch": 0.1640530759951749,
|
|
"grad_norm": 6.982624053955078,
|
|
"learning_rate": 8.225806451612904e-06,
|
|
"loss": 0.0772,
|
|
"step": 51
|
|
},
|
|
{
|
|
"epoch": 0.16726980297547245,
|
|
"grad_norm": 4.082429885864258,
|
|
"learning_rate": 8.387096774193549e-06,
|
|
"loss": 0.0734,
|
|
"step": 52
|
|
},
|
|
{
|
|
"epoch": 0.17048652995577002,
|
|
"grad_norm": 1.164082646369934,
|
|
"learning_rate": 8.548387096774194e-06,
|
|
"loss": 0.0541,
|
|
"step": 53
|
|
},
|
|
{
|
|
"epoch": 0.17370325693606756,
|
|
"grad_norm": 2.810659646987915,
|
|
"learning_rate": 8.70967741935484e-06,
|
|
"loss": 0.052,
|
|
"step": 54
|
|
},
|
|
{
|
|
"epoch": 0.1769199839163651,
|
|
"grad_norm": 1.5780754089355469,
|
|
"learning_rate": 8.870967741935484e-06,
|
|
"loss": 0.0582,
|
|
"step": 55
|
|
},
|
|
{
|
|
"epoch": 0.18013671089666264,
|
|
"grad_norm": 1.068620204925537,
|
|
"learning_rate": 9.03225806451613e-06,
|
|
"loss": 0.065,
|
|
"step": 56
|
|
},
|
|
{
|
|
"epoch": 0.18335343787696018,
|
|
"grad_norm": 2.523627758026123,
|
|
"learning_rate": 9.193548387096775e-06,
|
|
"loss": 0.0624,
|
|
"step": 57
|
|
},
|
|
{
|
|
"epoch": 0.18657016485725775,
|
|
"grad_norm": 1.244511365890503,
|
|
"learning_rate": 9.35483870967742e-06,
|
|
"loss": 0.0503,
|
|
"step": 58
|
|
},
|
|
{
|
|
"epoch": 0.1897868918375553,
|
|
"grad_norm": 0.9401457905769348,
|
|
"learning_rate": 9.516129032258065e-06,
|
|
"loss": 0.051,
|
|
"step": 59
|
|
},
|
|
{
|
|
"epoch": 0.19300361881785283,
|
|
"grad_norm": 0.9283802509307861,
|
|
"learning_rate": 9.67741935483871e-06,
|
|
"loss": 0.0537,
|
|
"step": 60
|
|
},
|
|
{
|
|
"epoch": 0.19622034579815037,
|
|
"grad_norm": 1.3737329244613647,
|
|
"learning_rate": 9.838709677419356e-06,
|
|
"loss": 0.0528,
|
|
"step": 61
|
|
},
|
|
{
|
|
"epoch": 0.19943707277844794,
|
|
"grad_norm": 0.8466247916221619,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.0505,
|
|
"step": 62
|
|
},
|
|
{
|
|
"epoch": 0.20265379975874548,
|
|
"grad_norm": 3.419365644454956,
|
|
"learning_rate": 9.999920755303033e-06,
|
|
"loss": 0.0696,
|
|
"step": 63
|
|
},
|
|
{
|
|
"epoch": 0.20587052673904302,
|
|
"grad_norm": 2.1508781909942627,
|
|
"learning_rate": 9.999683023724021e-06,
|
|
"loss": 0.0502,
|
|
"step": 64
|
|
},
|
|
{
|
|
"epoch": 0.20908725371934056,
|
|
"grad_norm": 1.2374366521835327,
|
|
"learning_rate": 9.99928681279855e-06,
|
|
"loss": 0.0536,
|
|
"step": 65
|
|
},
|
|
{
|
|
"epoch": 0.21230398069963813,
|
|
"grad_norm": 2.0750317573547363,
|
|
"learning_rate": 9.998732135085665e-06,
|
|
"loss": 0.0634,
|
|
"step": 66
|
|
},
|
|
{
|
|
"epoch": 0.21552070767993567,
|
|
"grad_norm": 1.1349234580993652,
|
|
"learning_rate": 9.998019008167476e-06,
|
|
"loss": 0.0472,
|
|
"step": 67
|
|
},
|
|
{
|
|
"epoch": 0.2187374346602332,
|
|
"grad_norm": 4.443458080291748,
|
|
"learning_rate": 9.99714745464859e-06,
|
|
"loss": 0.0714,
|
|
"step": 68
|
|
},
|
|
{
|
|
"epoch": 0.22195416164053075,
|
|
"grad_norm": 5.246551036834717,
|
|
"learning_rate": 9.99611750215541e-06,
|
|
"loss": 0.0686,
|
|
"step": 69
|
|
},
|
|
{
|
|
"epoch": 0.22517088862082832,
|
|
"grad_norm": 1.8846714496612549,
|
|
"learning_rate": 9.994929183335237e-06,
|
|
"loss": 0.0707,
|
|
"step": 70
|
|
},
|
|
{
|
|
"epoch": 0.22838761560112586,
|
|
"grad_norm": 2.367842674255371,
|
|
"learning_rate": 9.993582535855265e-06,
|
|
"loss": 0.0596,
|
|
"step": 71
|
|
},
|
|
{
|
|
"epoch": 0.2316043425814234,
|
|
"grad_norm": 3.4915974140167236,
|
|
"learning_rate": 9.992077602401358e-06,
|
|
"loss": 0.0644,
|
|
"step": 72
|
|
},
|
|
{
|
|
"epoch": 0.23482106956172094,
|
|
"grad_norm": 1.2369784116744995,
|
|
"learning_rate": 9.990414430676716e-06,
|
|
"loss": 0.0566,
|
|
"step": 73
|
|
},
|
|
{
|
|
"epoch": 0.2380377965420185,
|
|
"grad_norm": 1.0058597326278687,
|
|
"learning_rate": 9.988593073400354e-06,
|
|
"loss": 0.0518,
|
|
"step": 74
|
|
},
|
|
{
|
|
"epoch": 0.24125452352231605,
|
|
"grad_norm": 1.170558214187622,
|
|
"learning_rate": 9.986613588305435e-06,
|
|
"loss": 0.0481,
|
|
"step": 75
|
|
},
|
|
{
|
|
"epoch": 0.2444712505026136,
|
|
"grad_norm": 0.9504396915435791,
|
|
"learning_rate": 9.984476038137437e-06,
|
|
"loss": 0.0458,
|
|
"step": 76
|
|
},
|
|
{
|
|
"epoch": 0.24768797748291113,
|
|
"grad_norm": 4.625006675720215,
|
|
"learning_rate": 9.982180490652165e-06,
|
|
"loss": 0.0663,
|
|
"step": 77
|
|
},
|
|
{
|
|
"epoch": 0.25090470446320867,
|
|
"grad_norm": 2.3191418647766113,
|
|
"learning_rate": 9.979727018613607e-06,
|
|
"loss": 0.0439,
|
|
"step": 78
|
|
},
|
|
{
|
|
"epoch": 0.25412143144350624,
|
|
"grad_norm": 1.6738172769546509,
|
|
"learning_rate": 9.977115699791622e-06,
|
|
"loss": 0.0668,
|
|
"step": 79
|
|
},
|
|
{
|
|
"epoch": 0.25733815842380375,
|
|
"grad_norm": 2.8376853466033936,
|
|
"learning_rate": 9.974346616959476e-06,
|
|
"loss": 0.0612,
|
|
"step": 80
|
|
},
|
|
{
|
|
"epoch": 0.2605548854041013,
|
|
"grad_norm": 4.625660419464111,
|
|
"learning_rate": 9.971419857891223e-06,
|
|
"loss": 0.0602,
|
|
"step": 81
|
|
},
|
|
{
|
|
"epoch": 0.2637716123843989,
|
|
"grad_norm": 6.061123847961426,
|
|
"learning_rate": 9.968335515358916e-06,
|
|
"loss": 0.0623,
|
|
"step": 82
|
|
},
|
|
{
|
|
"epoch": 0.2669883393646964,
|
|
"grad_norm": 1.5065197944641113,
|
|
"learning_rate": 9.965093687129669e-06,
|
|
"loss": 0.0611,
|
|
"step": 83
|
|
},
|
|
{
|
|
"epoch": 0.27020506634499397,
|
|
"grad_norm": 0.9030428528785706,
|
|
"learning_rate": 9.961694475962562e-06,
|
|
"loss": 0.0623,
|
|
"step": 84
|
|
},
|
|
{
|
|
"epoch": 0.27342179332529154,
|
|
"grad_norm": 5.082237243652344,
|
|
"learning_rate": 9.95813798960538e-06,
|
|
"loss": 0.0758,
|
|
"step": 85
|
|
},
|
|
{
|
|
"epoch": 0.27663852030558905,
|
|
"grad_norm": 5.320428848266602,
|
|
"learning_rate": 9.954424340791195e-06,
|
|
"loss": 0.0675,
|
|
"step": 86
|
|
},
|
|
{
|
|
"epoch": 0.2798552472858866,
|
|
"grad_norm": 6.164220809936523,
|
|
"learning_rate": 9.950553647234798e-06,
|
|
"loss": 0.0713,
|
|
"step": 87
|
|
},
|
|
{
|
|
"epoch": 0.28307197426618413,
|
|
"grad_norm": 3.148167371749878,
|
|
"learning_rate": 9.94652603162896e-06,
|
|
"loss": 0.0577,
|
|
"step": 88
|
|
},
|
|
{
|
|
"epoch": 0.2862887012464817,
|
|
"grad_norm": 1.503680944442749,
|
|
"learning_rate": 9.942341621640558e-06,
|
|
"loss": 0.062,
|
|
"step": 89
|
|
},
|
|
{
|
|
"epoch": 0.28950542822677927,
|
|
"grad_norm": 2.4841108322143555,
|
|
"learning_rate": 9.938000549906509e-06,
|
|
"loss": 0.0594,
|
|
"step": 90
|
|
},
|
|
{
|
|
"epoch": 0.2927221552070768,
|
|
"grad_norm": 4.777047157287598,
|
|
"learning_rate": 9.93350295402958e-06,
|
|
"loss": 0.0601,
|
|
"step": 91
|
|
},
|
|
{
|
|
"epoch": 0.29593888218737435,
|
|
"grad_norm": 4.280494213104248,
|
|
"learning_rate": 9.92884897657402e-06,
|
|
"loss": 0.0654,
|
|
"step": 92
|
|
},
|
|
{
|
|
"epoch": 0.2991556091676719,
|
|
"grad_norm": 3.8119473457336426,
|
|
"learning_rate": 9.924038765061042e-06,
|
|
"loss": 0.0645,
|
|
"step": 93
|
|
},
|
|
{
|
|
"epoch": 0.30237233614796943,
|
|
"grad_norm": 1.2760810852050781,
|
|
"learning_rate": 9.919072471964146e-06,
|
|
"loss": 0.0502,
|
|
"step": 94
|
|
},
|
|
{
|
|
"epoch": 0.305589063128267,
|
|
"grad_norm": 2.0416059494018555,
|
|
"learning_rate": 9.913950254704291e-06,
|
|
"loss": 0.0601,
|
|
"step": 95
|
|
},
|
|
{
|
|
"epoch": 0.3088057901085645,
|
|
"grad_norm": 2.9436042308807373,
|
|
"learning_rate": 9.908672275644898e-06,
|
|
"loss": 0.0552,
|
|
"step": 96
|
|
},
|
|
{
|
|
"epoch": 0.3120225170888621,
|
|
"grad_norm": 2.80222487449646,
|
|
"learning_rate": 9.903238702086707e-06,
|
|
"loss": 0.0534,
|
|
"step": 97
|
|
},
|
|
{
|
|
"epoch": 0.31523924406915965,
|
|
"grad_norm": 2.396406888961792,
|
|
"learning_rate": 9.897649706262474e-06,
|
|
"loss": 0.0516,
|
|
"step": 98
|
|
},
|
|
{
|
|
"epoch": 0.31845597104945716,
|
|
"grad_norm": 1.1176875829696655,
|
|
"learning_rate": 9.89190546533151e-06,
|
|
"loss": 0.0544,
|
|
"step": 99
|
|
},
|
|
{
|
|
"epoch": 0.32167269802975473,
|
|
"grad_norm": 2.592433452606201,
|
|
"learning_rate": 9.88600616137407e-06,
|
|
"loss": 0.0493,
|
|
"step": 100
|
|
},
|
|
{
|
|
"epoch": 0.32488942501005225,
|
|
"grad_norm": 2.838442802429199,
|
|
"learning_rate": 9.879951981385577e-06,
|
|
"loss": 0.0628,
|
|
"step": 101
|
|
},
|
|
{
|
|
"epoch": 0.3281061519903498,
|
|
"grad_norm": 3.090041399002075,
|
|
"learning_rate": 9.873743117270691e-06,
|
|
"loss": 0.0426,
|
|
"step": 102
|
|
},
|
|
{
|
|
"epoch": 0.3313228789706474,
|
|
"grad_norm": 3.3808889389038086,
|
|
"learning_rate": 9.867379765837237e-06,
|
|
"loss": 0.0579,
|
|
"step": 103
|
|
},
|
|
{
|
|
"epoch": 0.3345396059509449,
|
|
"grad_norm": 0.756629228591919,
|
|
"learning_rate": 9.860862128789954e-06,
|
|
"loss": 0.0537,
|
|
"step": 104
|
|
},
|
|
{
|
|
"epoch": 0.33775633293124246,
|
|
"grad_norm": 1.3207106590270996,
|
|
"learning_rate": 9.854190412724114e-06,
|
|
"loss": 0.0406,
|
|
"step": 105
|
|
},
|
|
{
|
|
"epoch": 0.34097305991154003,
|
|
"grad_norm": 2.726410388946533,
|
|
"learning_rate": 9.847364829118963e-06,
|
|
"loss": 0.0603,
|
|
"step": 106
|
|
},
|
|
{
|
|
"epoch": 0.34418978689183755,
|
|
"grad_norm": 1.6553726196289062,
|
|
"learning_rate": 9.840385594331022e-06,
|
|
"loss": 0.0635,
|
|
"step": 107
|
|
},
|
|
{
|
|
"epoch": 0.3474065138721351,
|
|
"grad_norm": 0.6768801808357239,
|
|
"learning_rate": 9.833252929587231e-06,
|
|
"loss": 0.0538,
|
|
"step": 108
|
|
},
|
|
{
|
|
"epoch": 0.3506232408524326,
|
|
"grad_norm": 2.1218836307525635,
|
|
"learning_rate": 9.825967060977933e-06,
|
|
"loss": 0.0503,
|
|
"step": 109
|
|
},
|
|
{
|
|
"epoch": 0.3538399678327302,
|
|
"grad_norm": 2.7606120109558105,
|
|
"learning_rate": 9.818528219449705e-06,
|
|
"loss": 0.05,
|
|
"step": 110
|
|
},
|
|
{
|
|
"epoch": 0.35705669481302776,
|
|
"grad_norm": 1.1629998683929443,
|
|
"learning_rate": 9.810936640798046e-06,
|
|
"loss": 0.0456,
|
|
"step": 111
|
|
},
|
|
{
|
|
"epoch": 0.3602734217933253,
|
|
"grad_norm": 1.027559757232666,
|
|
"learning_rate": 9.803192565659898e-06,
|
|
"loss": 0.0522,
|
|
"step": 112
|
|
},
|
|
{
|
|
"epoch": 0.36349014877362285,
|
|
"grad_norm": 0.8349726796150208,
|
|
"learning_rate": 9.795296239506011e-06,
|
|
"loss": 0.0424,
|
|
"step": 113
|
|
},
|
|
{
|
|
"epoch": 0.36670687575392036,
|
|
"grad_norm": 1.19048273563385,
|
|
"learning_rate": 9.78724791263318e-06,
|
|
"loss": 0.0472,
|
|
"step": 114
|
|
},
|
|
{
|
|
"epoch": 0.3699236027342179,
|
|
"grad_norm": 0.9269819259643555,
|
|
"learning_rate": 9.779047840156288e-06,
|
|
"loss": 0.0528,
|
|
"step": 115
|
|
},
|
|
{
|
|
"epoch": 0.3731403297145155,
|
|
"grad_norm": 0.7386018633842468,
|
|
"learning_rate": 9.770696282000245e-06,
|
|
"loss": 0.0561,
|
|
"step": 116
|
|
},
|
|
{
|
|
"epoch": 0.376357056694813,
|
|
"grad_norm": 0.8170301914215088,
|
|
"learning_rate": 9.762193502891726e-06,
|
|
"loss": 0.0457,
|
|
"step": 117
|
|
},
|
|
{
|
|
"epoch": 0.3795737836751106,
|
|
"grad_norm": 1.0884501934051514,
|
|
"learning_rate": 9.753539772350792e-06,
|
|
"loss": 0.0467,
|
|
"step": 118
|
|
},
|
|
{
|
|
"epoch": 0.38279051065540814,
|
|
"grad_norm": 1.1235942840576172,
|
|
"learning_rate": 9.744735364682347e-06,
|
|
"loss": 0.0406,
|
|
"step": 119
|
|
},
|
|
{
|
|
"epoch": 0.38600723763570566,
|
|
"grad_norm": 0.8838767409324646,
|
|
"learning_rate": 9.735780558967434e-06,
|
|
"loss": 0.0487,
|
|
"step": 120
|
|
},
|
|
{
|
|
"epoch": 0.3892239646160032,
|
|
"grad_norm": 0.9004449248313904,
|
|
"learning_rate": 9.726675639054403e-06,
|
|
"loss": 0.0543,
|
|
"step": 121
|
|
},
|
|
{
|
|
"epoch": 0.39244069159630074,
|
|
"grad_norm": 1.97210693359375,
|
|
"learning_rate": 9.717420893549902e-06,
|
|
"loss": 0.0493,
|
|
"step": 122
|
|
},
|
|
{
|
|
"epoch": 0.3956574185765983,
|
|
"grad_norm": 0.7708803415298462,
|
|
"learning_rate": 9.70801661580973e-06,
|
|
"loss": 0.0437,
|
|
"step": 123
|
|
},
|
|
{
|
|
"epoch": 0.3988741455568959,
|
|
"grad_norm": 1.6689863204956055,
|
|
"learning_rate": 9.698463103929542e-06,
|
|
"loss": 0.0455,
|
|
"step": 124
|
|
},
|
|
{
|
|
"epoch": 0.4020908725371934,
|
|
"grad_norm": 1.5968708992004395,
|
|
"learning_rate": 9.688760660735403e-06,
|
|
"loss": 0.0569,
|
|
"step": 125
|
|
},
|
|
{
|
|
"epoch": 0.40530759951749096,
|
|
"grad_norm": 1.1041440963745117,
|
|
"learning_rate": 9.67890959377418e-06,
|
|
"loss": 0.0576,
|
|
"step": 126
|
|
},
|
|
{
|
|
"epoch": 0.4085243264977885,
|
|
"grad_norm": 2.1897428035736084,
|
|
"learning_rate": 9.668910215303797e-06,
|
|
"loss": 0.0514,
|
|
"step": 127
|
|
},
|
|
{
|
|
"epoch": 0.41174105347808604,
|
|
"grad_norm": 1.594397783279419,
|
|
"learning_rate": 9.658762842283343e-06,
|
|
"loss": 0.0444,
|
|
"step": 128
|
|
},
|
|
{
|
|
"epoch": 0.4149577804583836,
|
|
"grad_norm": 1.404996395111084,
|
|
"learning_rate": 9.648467796363019e-06,
|
|
"loss": 0.0493,
|
|
"step": 129
|
|
},
|
|
{
|
|
"epoch": 0.4181745074386811,
|
|
"grad_norm": 0.889514148235321,
|
|
"learning_rate": 9.638025403873939e-06,
|
|
"loss": 0.0576,
|
|
"step": 130
|
|
},
|
|
{
|
|
"epoch": 0.4213912344189787,
|
|
"grad_norm": 3.748598337173462,
|
|
"learning_rate": 9.627435995817799e-06,
|
|
"loss": 0.0599,
|
|
"step": 131
|
|
},
|
|
{
|
|
"epoch": 0.42460796139927626,
|
|
"grad_norm": 4.13529109954834,
|
|
"learning_rate": 9.616699907856368e-06,
|
|
"loss": 0.0625,
|
|
"step": 132
|
|
},
|
|
{
|
|
"epoch": 0.42782468837957377,
|
|
"grad_norm": 2.6388087272644043,
|
|
"learning_rate": 9.605817480300863e-06,
|
|
"loss": 0.0536,
|
|
"step": 133
|
|
},
|
|
{
|
|
"epoch": 0.43104141535987134,
|
|
"grad_norm": 1.5752240419387817,
|
|
"learning_rate": 9.594789058101154e-06,
|
|
"loss": 0.0636,
|
|
"step": 134
|
|
},
|
|
{
|
|
"epoch": 0.43425814234016885,
|
|
"grad_norm": 1.7247588634490967,
|
|
"learning_rate": 9.58361499083483e-06,
|
|
"loss": 0.0602,
|
|
"step": 135
|
|
},
|
|
{
|
|
"epoch": 0.4374748693204664,
|
|
"grad_norm": 1.2270214557647705,
|
|
"learning_rate": 9.57229563269612e-06,
|
|
"loss": 0.0543,
|
|
"step": 136
|
|
},
|
|
{
|
|
"epoch": 0.440691596300764,
|
|
"grad_norm": 1.9172810316085815,
|
|
"learning_rate": 9.560831342484668e-06,
|
|
"loss": 0.0456,
|
|
"step": 137
|
|
},
|
|
{
|
|
"epoch": 0.4439083232810615,
|
|
"grad_norm": 5.208808898925781,
|
|
"learning_rate": 9.549222483594154e-06,
|
|
"loss": 0.0868,
|
|
"step": 138
|
|
},
|
|
{
|
|
"epoch": 0.44712505026135907,
|
|
"grad_norm": 4.008519172668457,
|
|
"learning_rate": 9.53746942400078e-06,
|
|
"loss": 0.0646,
|
|
"step": 139
|
|
},
|
|
{
|
|
"epoch": 0.45034177724165664,
|
|
"grad_norm": 2.6669585704803467,
|
|
"learning_rate": 9.525572536251608e-06,
|
|
"loss": 0.068,
|
|
"step": 140
|
|
},
|
|
{
|
|
"epoch": 0.45355850422195415,
|
|
"grad_norm": 0.6555745601654053,
|
|
"learning_rate": 9.513532197452737e-06,
|
|
"loss": 0.0539,
|
|
"step": 141
|
|
},
|
|
{
|
|
"epoch": 0.4567752312022517,
|
|
"grad_norm": 3.599552869796753,
|
|
"learning_rate": 9.501348789257373e-06,
|
|
"loss": 0.0515,
|
|
"step": 142
|
|
},
|
|
{
|
|
"epoch": 0.45999195818254923,
|
|
"grad_norm": 5.188879489898682,
|
|
"learning_rate": 9.48902269785371e-06,
|
|
"loss": 0.0754,
|
|
"step": 143
|
|
},
|
|
{
|
|
"epoch": 0.4632086851628468,
|
|
"grad_norm": 3.3630921840667725,
|
|
"learning_rate": 9.476554313952697e-06,
|
|
"loss": 0.0651,
|
|
"step": 144
|
|
},
|
|
{
|
|
"epoch": 0.46642541214314437,
|
|
"grad_norm": 2.4736194610595703,
|
|
"learning_rate": 9.46394403277566e-06,
|
|
"loss": 0.0578,
|
|
"step": 145
|
|
},
|
|
{
|
|
"epoch": 0.4696421391234419,
|
|
"grad_norm": 0.7932046055793762,
|
|
"learning_rate": 9.451192254041759e-06,
|
|
"loss": 0.0466,
|
|
"step": 146
|
|
},
|
|
{
|
|
"epoch": 0.47285886610373945,
|
|
"grad_norm": 0.980038583278656,
|
|
"learning_rate": 9.438299381955333e-06,
|
|
"loss": 0.0452,
|
|
"step": 147
|
|
},
|
|
{
|
|
"epoch": 0.476075593084037,
|
|
"grad_norm": 1.6979252099990845,
|
|
"learning_rate": 9.425265825193077e-06,
|
|
"loss": 0.0445,
|
|
"step": 148
|
|
},
|
|
{
|
|
"epoch": 0.47929232006433453,
|
|
"grad_norm": 2.8460330963134766,
|
|
"learning_rate": 9.412091996891097e-06,
|
|
"loss": 0.051,
|
|
"step": 149
|
|
},
|
|
{
|
|
"epoch": 0.4825090470446321,
|
|
"grad_norm": 1.3673274517059326,
|
|
"learning_rate": 9.398778314631801e-06,
|
|
"loss": 0.0473,
|
|
"step": 150
|
|
},
|
|
{
|
|
"epoch": 0.4857257740249296,
|
|
"grad_norm": 1.6840636730194092,
|
|
"learning_rate": 9.385325200430679e-06,
|
|
"loss": 0.0525,
|
|
"step": 151
|
|
},
|
|
{
|
|
"epoch": 0.4889425010052272,
|
|
"grad_norm": 2.1259257793426514,
|
|
"learning_rate": 9.371733080722911e-06,
|
|
"loss": 0.0548,
|
|
"step": 152
|
|
},
|
|
{
|
|
"epoch": 0.49215922798552475,
|
|
"grad_norm": 0.7495786547660828,
|
|
"learning_rate": 9.358002386349862e-06,
|
|
"loss": 0.0379,
|
|
"step": 153
|
|
},
|
|
{
|
|
"epoch": 0.49537595496582226,
|
|
"grad_norm": 1.248661756515503,
|
|
"learning_rate": 9.34413355254542e-06,
|
|
"loss": 0.0477,
|
|
"step": 154
|
|
},
|
|
{
|
|
"epoch": 0.49859268194611983,
|
|
"grad_norm": 1.250671148300171,
|
|
"learning_rate": 9.330127018922195e-06,
|
|
"loss": 0.0524,
|
|
"step": 155
|
|
},
|
|
{
|
|
"epoch": 0.5018094089264173,
|
|
"grad_norm": 0.5698635578155518,
|
|
"learning_rate": 9.31598322945759e-06,
|
|
"loss": 0.0508,
|
|
"step": 156
|
|
},
|
|
{
|
|
"epoch": 0.505026135906715,
|
|
"grad_norm": 0.6476792097091675,
|
|
"learning_rate": 9.301702632479734e-06,
|
|
"loss": 0.0482,
|
|
"step": 157
|
|
},
|
|
{
|
|
"epoch": 0.5082428628870125,
|
|
"grad_norm": 0.6281876564025879,
|
|
"learning_rate": 9.287285680653254e-06,
|
|
"loss": 0.0448,
|
|
"step": 158
|
|
},
|
|
{
|
|
"epoch": 0.51145958986731,
|
|
"grad_norm": 0.8750901222229004,
|
|
"learning_rate": 9.272732830964948e-06,
|
|
"loss": 0.0652,
|
|
"step": 159
|
|
},
|
|
{
|
|
"epoch": 0.5146763168476075,
|
|
"grad_norm": 0.6696067452430725,
|
|
"learning_rate": 9.258044544709276e-06,
|
|
"loss": 0.0432,
|
|
"step": 160
|
|
},
|
|
{
|
|
"epoch": 0.5178930438279051,
|
|
"grad_norm": 0.5400009155273438,
|
|
"learning_rate": 9.243221287473755e-06,
|
|
"loss": 0.0389,
|
|
"step": 161
|
|
},
|
|
{
|
|
"epoch": 0.5211097708082026,
|
|
"grad_norm": 0.59368497133255,
|
|
"learning_rate": 9.228263529124199e-06,
|
|
"loss": 0.0407,
|
|
"step": 162
|
|
},
|
|
{
|
|
"epoch": 0.5243264977885002,
|
|
"grad_norm": 0.8467090725898743,
|
|
"learning_rate": 9.21317174378982e-06,
|
|
"loss": 0.0489,
|
|
"step": 163
|
|
},
|
|
{
|
|
"epoch": 0.5275432247687978,
|
|
"grad_norm": 0.6505163908004761,
|
|
"learning_rate": 9.197946409848196e-06,
|
|
"loss": 0.0448,
|
|
"step": 164
|
|
},
|
|
{
|
|
"epoch": 0.5307599517490953,
|
|
"grad_norm": 1.452300786972046,
|
|
"learning_rate": 9.182588009910119e-06,
|
|
"loss": 0.0487,
|
|
"step": 165
|
|
},
|
|
{
|
|
"epoch": 0.5339766787293928,
|
|
"grad_norm": 1.1281909942626953,
|
|
"learning_rate": 9.167097030804289e-06,
|
|
"loss": 0.0468,
|
|
"step": 166
|
|
},
|
|
{
|
|
"epoch": 0.5371934057096904,
|
|
"grad_norm": 1.3461655378341675,
|
|
"learning_rate": 9.151473963561884e-06,
|
|
"loss": 0.0589,
|
|
"step": 167
|
|
},
|
|
{
|
|
"epoch": 0.5404101326899879,
|
|
"grad_norm": 0.5634208917617798,
|
|
"learning_rate": 9.135719303400995e-06,
|
|
"loss": 0.0405,
|
|
"step": 168
|
|
},
|
|
{
|
|
"epoch": 0.5436268596702855,
|
|
"grad_norm": 1.198769450187683,
|
|
"learning_rate": 9.119833549710927e-06,
|
|
"loss": 0.0635,
|
|
"step": 169
|
|
},
|
|
{
|
|
"epoch": 0.5468435866505831,
|
|
"grad_norm": 1.4971657991409302,
|
|
"learning_rate": 9.103817206036383e-06,
|
|
"loss": 0.05,
|
|
"step": 170
|
|
},
|
|
{
|
|
"epoch": 0.5500603136308806,
|
|
"grad_norm": 0.9563581347465515,
|
|
"learning_rate": 9.087670780061477e-06,
|
|
"loss": 0.0555,
|
|
"step": 171
|
|
},
|
|
{
|
|
"epoch": 0.5532770406111781,
|
|
"grad_norm": 0.7893424034118652,
|
|
"learning_rate": 9.071394783593664e-06,
|
|
"loss": 0.0495,
|
|
"step": 172
|
|
},
|
|
{
|
|
"epoch": 0.5564937675914756,
|
|
"grad_norm": 1.9164668321609497,
|
|
"learning_rate": 9.054989732547507e-06,
|
|
"loss": 0.0519,
|
|
"step": 173
|
|
},
|
|
{
|
|
"epoch": 0.5597104945717732,
|
|
"grad_norm": 0.7731283903121948,
|
|
"learning_rate": 9.038456146928325e-06,
|
|
"loss": 0.047,
|
|
"step": 174
|
|
},
|
|
{
|
|
"epoch": 0.5629272215520708,
|
|
"grad_norm": 1.0580449104309082,
|
|
"learning_rate": 9.021794550815713e-06,
|
|
"loss": 0.052,
|
|
"step": 175
|
|
},
|
|
{
|
|
"epoch": 0.5661439485323683,
|
|
"grad_norm": 0.7994678616523743,
|
|
"learning_rate": 9.005005472346923e-06,
|
|
"loss": 0.0587,
|
|
"step": 176
|
|
},
|
|
{
|
|
"epoch": 0.5693606755126659,
|
|
"grad_norm": 1.5067932605743408,
|
|
"learning_rate": 8.988089443700131e-06,
|
|
"loss": 0.0426,
|
|
"step": 177
|
|
},
|
|
{
|
|
"epoch": 0.5725774024929634,
|
|
"grad_norm": 1.0666264295578003,
|
|
"learning_rate": 8.971047001077561e-06,
|
|
"loss": 0.0432,
|
|
"step": 178
|
|
},
|
|
{
|
|
"epoch": 0.5757941294732609,
|
|
"grad_norm": 2.072291135787964,
|
|
"learning_rate": 8.953878684688492e-06,
|
|
"loss": 0.0467,
|
|
"step": 179
|
|
},
|
|
{
|
|
"epoch": 0.5790108564535585,
|
|
"grad_norm": 0.4865788221359253,
|
|
"learning_rate": 8.936585038732143e-06,
|
|
"loss": 0.0375,
|
|
"step": 180
|
|
},
|
|
{
|
|
"epoch": 0.582227583433856,
|
|
"grad_norm": 2.6773552894592285,
|
|
"learning_rate": 8.919166611380397e-06,
|
|
"loss": 0.065,
|
|
"step": 181
|
|
},
|
|
{
|
|
"epoch": 0.5854443104141536,
|
|
"grad_norm": 3.049079656600952,
|
|
"learning_rate": 8.90162395476046e-06,
|
|
"loss": 0.0652,
|
|
"step": 182
|
|
},
|
|
{
|
|
"epoch": 0.5886610373944512,
|
|
"grad_norm": 1.1698511838912964,
|
|
"learning_rate": 8.883957624937333e-06,
|
|
"loss": 0.0519,
|
|
"step": 183
|
|
},
|
|
{
|
|
"epoch": 0.5918777643747487,
|
|
"grad_norm": 1.1253437995910645,
|
|
"learning_rate": 8.866168181896198e-06,
|
|
"loss": 0.0562,
|
|
"step": 184
|
|
},
|
|
{
|
|
"epoch": 0.5950944913550462,
|
|
"grad_norm": 0.5882744789123535,
|
|
"learning_rate": 8.848256189524661e-06,
|
|
"loss": 0.0472,
|
|
"step": 185
|
|
},
|
|
{
|
|
"epoch": 0.5983112183353438,
|
|
"grad_norm": 0.5303323864936829,
|
|
"learning_rate": 8.83022221559489e-06,
|
|
"loss": 0.0359,
|
|
"step": 186
|
|
},
|
|
{
|
|
"epoch": 0.6015279453156414,
|
|
"grad_norm": 0.8436638712882996,
|
|
"learning_rate": 8.812066831745602e-06,
|
|
"loss": 0.0335,
|
|
"step": 187
|
|
},
|
|
{
|
|
"epoch": 0.6047446722959389,
|
|
"grad_norm": 0.7851281762123108,
|
|
"learning_rate": 8.793790613463956e-06,
|
|
"loss": 0.0459,
|
|
"step": 188
|
|
},
|
|
{
|
|
"epoch": 0.6079613992762364,
|
|
"grad_norm": 0.7391752600669861,
|
|
"learning_rate": 8.775394140067299e-06,
|
|
"loss": 0.0445,
|
|
"step": 189
|
|
},
|
|
{
|
|
"epoch": 0.611178126256534,
|
|
"grad_norm": 0.9071338772773743,
|
|
"learning_rate": 8.756877994684818e-06,
|
|
"loss": 0.0468,
|
|
"step": 190
|
|
},
|
|
{
|
|
"epoch": 0.6143948532368315,
|
|
"grad_norm": 1.225618600845337,
|
|
"learning_rate": 8.738242764239046e-06,
|
|
"loss": 0.0433,
|
|
"step": 191
|
|
},
|
|
{
|
|
"epoch": 0.617611580217129,
|
|
"grad_norm": 1.5072846412658691,
|
|
"learning_rate": 8.719489039427256e-06,
|
|
"loss": 0.0467,
|
|
"step": 192
|
|
},
|
|
{
|
|
"epoch": 0.6208283071974267,
|
|
"grad_norm": 0.7594591975212097,
|
|
"learning_rate": 8.700617414702746e-06,
|
|
"loss": 0.0407,
|
|
"step": 193
|
|
},
|
|
{
|
|
"epoch": 0.6240450341777242,
|
|
"grad_norm": 1.7316999435424805,
|
|
"learning_rate": 8.681628488255986e-06,
|
|
"loss": 0.0524,
|
|
"step": 194
|
|
},
|
|
{
|
|
"epoch": 0.6272617611580217,
|
|
"grad_norm": 0.6949440240859985,
|
|
"learning_rate": 8.66252286199567e-06,
|
|
"loss": 0.0354,
|
|
"step": 195
|
|
},
|
|
{
|
|
"epoch": 0.6304784881383193,
|
|
"grad_norm": 0.6324911117553711,
|
|
"learning_rate": 8.643301141529619e-06,
|
|
"loss": 0.0525,
|
|
"step": 196
|
|
},
|
|
{
|
|
"epoch": 0.6336952151186168,
|
|
"grad_norm": 0.6617627739906311,
|
|
"learning_rate": 8.6239639361456e-06,
|
|
"loss": 0.0471,
|
|
"step": 197
|
|
},
|
|
{
|
|
"epoch": 0.6369119420989143,
|
|
"grad_norm": 1.0735185146331787,
|
|
"learning_rate": 8.604511858792006e-06,
|
|
"loss": 0.0347,
|
|
"step": 198
|
|
},
|
|
{
|
|
"epoch": 0.640128669079212,
|
|
"grad_norm": 0.6503901481628418,
|
|
"learning_rate": 8.584945526058426e-06,
|
|
"loss": 0.0621,
|
|
"step": 199
|
|
},
|
|
{
|
|
"epoch": 0.6433453960595095,
|
|
"grad_norm": 0.5145460367202759,
|
|
"learning_rate": 8.565265558156101e-06,
|
|
"loss": 0.0444,
|
|
"step": 200
|
|
},
|
|
{
|
|
"epoch": 0.646562123039807,
|
|
"grad_norm": 0.5394482016563416,
|
|
"learning_rate": 8.545472578898276e-06,
|
|
"loss": 0.0505,
|
|
"step": 201
|
|
},
|
|
{
|
|
"epoch": 0.6497788500201045,
|
|
"grad_norm": 0.9032294750213623,
|
|
"learning_rate": 8.525567215680397e-06,
|
|
"loss": 0.0522,
|
|
"step": 202
|
|
},
|
|
{
|
|
"epoch": 0.6529955770004021,
|
|
"grad_norm": 0.4706067144870758,
|
|
"learning_rate": 8.505550099460264e-06,
|
|
"loss": 0.0388,
|
|
"step": 203
|
|
},
|
|
{
|
|
"epoch": 0.6562123039806996,
|
|
"grad_norm": 0.602165699005127,
|
|
"learning_rate": 8.485421864737997e-06,
|
|
"loss": 0.0557,
|
|
"step": 204
|
|
},
|
|
{
|
|
"epoch": 0.6594290309609971,
|
|
"grad_norm": 1.2543528079986572,
|
|
"learning_rate": 8.465183149535939e-06,
|
|
"loss": 0.0539,
|
|
"step": 205
|
|
},
|
|
{
|
|
"epoch": 0.6626457579412948,
|
|
"grad_norm": 1.1637508869171143,
|
|
"learning_rate": 8.444834595378434e-06,
|
|
"loss": 0.0526,
|
|
"step": 206
|
|
},
|
|
{
|
|
"epoch": 0.6658624849215923,
|
|
"grad_norm": 2.013043165206909,
|
|
"learning_rate": 8.424376847271483e-06,
|
|
"loss": 0.0523,
|
|
"step": 207
|
|
},
|
|
{
|
|
"epoch": 0.6690792119018898,
|
|
"grad_norm": 1.166269302368164,
|
|
"learning_rate": 8.403810553682307e-06,
|
|
"loss": 0.0482,
|
|
"step": 208
|
|
},
|
|
{
|
|
"epoch": 0.6722959388821874,
|
|
"grad_norm": 1.0098134279251099,
|
|
"learning_rate": 8.383136366518788e-06,
|
|
"loss": 0.0477,
|
|
"step": 209
|
|
},
|
|
{
|
|
"epoch": 0.6755126658624849,
|
|
"grad_norm": 0.7044484615325928,
|
|
"learning_rate": 8.362354941108803e-06,
|
|
"loss": 0.05,
|
|
"step": 210
|
|
},
|
|
{
|
|
"epoch": 0.6787293928427824,
|
|
"grad_norm": 0.5775832533836365,
|
|
"learning_rate": 8.341466936179457e-06,
|
|
"loss": 0.0532,
|
|
"step": 211
|
|
},
|
|
{
|
|
"epoch": 0.6819461198230801,
|
|
"grad_norm": 1.780887246131897,
|
|
"learning_rate": 8.320473013836197e-06,
|
|
"loss": 0.0385,
|
|
"step": 212
|
|
},
|
|
{
|
|
"epoch": 0.6851628468033776,
|
|
"grad_norm": 0.6344701051712036,
|
|
"learning_rate": 8.299373839541829e-06,
|
|
"loss": 0.0345,
|
|
"step": 213
|
|
},
|
|
{
|
|
"epoch": 0.6883795737836751,
|
|
"grad_norm": 0.597938597202301,
|
|
"learning_rate": 8.278170082095422e-06,
|
|
"loss": 0.0549,
|
|
"step": 214
|
|
},
|
|
{
|
|
"epoch": 0.6915963007639726,
|
|
"grad_norm": 0.5144065022468567,
|
|
"learning_rate": 8.256862413611113e-06,
|
|
"loss": 0.0368,
|
|
"step": 215
|
|
},
|
|
{
|
|
"epoch": 0.6948130277442702,
|
|
"grad_norm": 0.5459667444229126,
|
|
"learning_rate": 8.23545150949679e-06,
|
|
"loss": 0.0396,
|
|
"step": 216
|
|
},
|
|
{
|
|
"epoch": 0.6980297547245677,
|
|
"grad_norm": 1.465198278427124,
|
|
"learning_rate": 8.213938048432697e-06,
|
|
"loss": 0.0534,
|
|
"step": 217
|
|
},
|
|
{
|
|
"epoch": 0.7012464817048653,
|
|
"grad_norm": 0.9042430520057678,
|
|
"learning_rate": 8.192322712349917e-06,
|
|
"loss": 0.0463,
|
|
"step": 218
|
|
},
|
|
{
|
|
"epoch": 0.7044632086851629,
|
|
"grad_norm": 1.0576472282409668,
|
|
"learning_rate": 8.170606186408756e-06,
|
|
"loss": 0.0392,
|
|
"step": 219
|
|
},
|
|
{
|
|
"epoch": 0.7076799356654604,
|
|
"grad_norm": 0.7928400039672852,
|
|
"learning_rate": 8.148789158977012e-06,
|
|
"loss": 0.056,
|
|
"step": 220
|
|
},
|
|
{
|
|
"epoch": 0.7108966626457579,
|
|
"grad_norm": 1.0211706161499023,
|
|
"learning_rate": 8.126872321608185e-06,
|
|
"loss": 0.059,
|
|
"step": 221
|
|
},
|
|
{
|
|
"epoch": 0.7141133896260555,
|
|
"grad_norm": 0.5951483845710754,
|
|
"learning_rate": 8.104856369019525e-06,
|
|
"loss": 0.0493,
|
|
"step": 222
|
|
},
|
|
{
|
|
"epoch": 0.717330116606353,
|
|
"grad_norm": 1.3035422563552856,
|
|
"learning_rate": 8.08274199907003e-06,
|
|
"loss": 0.0411,
|
|
"step": 223
|
|
},
|
|
{
|
|
"epoch": 0.7205468435866506,
|
|
"grad_norm": 0.7945671081542969,
|
|
"learning_rate": 8.060529912738316e-06,
|
|
"loss": 0.0419,
|
|
"step": 224
|
|
},
|
|
{
|
|
"epoch": 0.7237635705669482,
|
|
"grad_norm": 0.6303293704986572,
|
|
"learning_rate": 8.038220814100403e-06,
|
|
"loss": 0.0504,
|
|
"step": 225
|
|
},
|
|
{
|
|
"epoch": 0.7269802975472457,
|
|
"grad_norm": 1.6979130506515503,
|
|
"learning_rate": 8.0158154103074e-06,
|
|
"loss": 0.0512,
|
|
"step": 226
|
|
},
|
|
{
|
|
"epoch": 0.7301970245275432,
|
|
"grad_norm": 0.7927567362785339,
|
|
"learning_rate": 7.993314411563075e-06,
|
|
"loss": 0.0559,
|
|
"step": 227
|
|
},
|
|
{
|
|
"epoch": 0.7334137515078407,
|
|
"grad_norm": 0.7262939214706421,
|
|
"learning_rate": 7.970718531101365e-06,
|
|
"loss": 0.0439,
|
|
"step": 228
|
|
},
|
|
{
|
|
"epoch": 0.7366304784881383,
|
|
"grad_norm": 0.7317245006561279,
|
|
"learning_rate": 7.948028485163744e-06,
|
|
"loss": 0.0399,
|
|
"step": 229
|
|
},
|
|
{
|
|
"epoch": 0.7398472054684359,
|
|
"grad_norm": 0.9586815237998962,
|
|
"learning_rate": 7.925244992976538e-06,
|
|
"loss": 0.0573,
|
|
"step": 230
|
|
},
|
|
{
|
|
"epoch": 0.7430639324487334,
|
|
"grad_norm": 1.5452810525894165,
|
|
"learning_rate": 7.902368776728125e-06,
|
|
"loss": 0.0464,
|
|
"step": 231
|
|
},
|
|
{
|
|
"epoch": 0.746280659429031,
|
|
"grad_norm": 0.8795072436332703,
|
|
"learning_rate": 7.879400561546033e-06,
|
|
"loss": 0.0453,
|
|
"step": 232
|
|
},
|
|
{
|
|
"epoch": 0.7494973864093285,
|
|
"grad_norm": 0.648480236530304,
|
|
"learning_rate": 7.856341075473963e-06,
|
|
"loss": 0.0467,
|
|
"step": 233
|
|
},
|
|
{
|
|
"epoch": 0.752714113389626,
|
|
"grad_norm": 0.646035373210907,
|
|
"learning_rate": 7.833191049448706e-06,
|
|
"loss": 0.0377,
|
|
"step": 234
|
|
},
|
|
{
|
|
"epoch": 0.7559308403699236,
|
|
"grad_norm": 1.1588225364685059,
|
|
"learning_rate": 7.809951217276986e-06,
|
|
"loss": 0.0427,
|
|
"step": 235
|
|
},
|
|
{
|
|
"epoch": 0.7591475673502212,
|
|
"grad_norm": 0.6012084484100342,
|
|
"learning_rate": 7.786622315612182e-06,
|
|
"loss": 0.0433,
|
|
"step": 236
|
|
},
|
|
{
|
|
"epoch": 0.7623642943305187,
|
|
"grad_norm": 0.5877156853675842,
|
|
"learning_rate": 7.763205083930995e-06,
|
|
"loss": 0.0394,
|
|
"step": 237
|
|
},
|
|
{
|
|
"epoch": 0.7655810213108163,
|
|
"grad_norm": 0.7161210179328918,
|
|
"learning_rate": 7.739700264509993e-06,
|
|
"loss": 0.0379,
|
|
"step": 238
|
|
},
|
|
{
|
|
"epoch": 0.7687977482911138,
|
|
"grad_norm": 0.819977343082428,
|
|
"learning_rate": 7.716108602402094e-06,
|
|
"loss": 0.0599,
|
|
"step": 239
|
|
},
|
|
{
|
|
"epoch": 0.7720144752714113,
|
|
"grad_norm": 0.6733528971672058,
|
|
"learning_rate": 7.692430845412946e-06,
|
|
"loss": 0.0465,
|
|
"step": 240
|
|
},
|
|
{
|
|
"epoch": 0.7752312022517089,
|
|
"grad_norm": 0.6228499412536621,
|
|
"learning_rate": 7.668667744077215e-06,
|
|
"loss": 0.0478,
|
|
"step": 241
|
|
},
|
|
{
|
|
"epoch": 0.7784479292320065,
|
|
"grad_norm": 0.8519704341888428,
|
|
"learning_rate": 7.644820051634813e-06,
|
|
"loss": 0.0503,
|
|
"step": 242
|
|
},
|
|
{
|
|
"epoch": 0.781664656212304,
|
|
"grad_norm": 0.966986358165741,
|
|
"learning_rate": 7.6208885240069995e-06,
|
|
"loss": 0.0412,
|
|
"step": 243
|
|
},
|
|
{
|
|
"epoch": 0.7848813831926015,
|
|
"grad_norm": 0.8694539070129395,
|
|
"learning_rate": 7.596873919772438e-06,
|
|
"loss": 0.0436,
|
|
"step": 244
|
|
},
|
|
{
|
|
"epoch": 0.7880981101728991,
|
|
"grad_norm": 0.8280279040336609,
|
|
"learning_rate": 7.572777000143145e-06,
|
|
"loss": 0.0347,
|
|
"step": 245
|
|
},
|
|
{
|
|
"epoch": 0.7913148371531966,
|
|
"grad_norm": 1.55968177318573,
|
|
"learning_rate": 7.548598528940354e-06,
|
|
"loss": 0.0478,
|
|
"step": 246
|
|
},
|
|
{
|
|
"epoch": 0.7945315641334941,
|
|
"grad_norm": 0.5930286049842834,
|
|
"learning_rate": 7.524339272570317e-06,
|
|
"loss": 0.0511,
|
|
"step": 247
|
|
},
|
|
{
|
|
"epoch": 0.7977482911137918,
|
|
"grad_norm": 1.2295900583267212,
|
|
"learning_rate": 7.500000000000001e-06,
|
|
"loss": 0.055,
|
|
"step": 248
|
|
},
|
|
{
|
|
"epoch": 0.8009650180940893,
|
|
"grad_norm": 1.3309237957000732,
|
|
"learning_rate": 7.475581482732717e-06,
|
|
"loss": 0.0557,
|
|
"step": 249
|
|
},
|
|
{
|
|
"epoch": 0.8041817450743868,
|
|
"grad_norm": 0.7122506499290466,
|
|
"learning_rate": 7.451084494783668e-06,
|
|
"loss": 0.0469,
|
|
"step": 250
|
|
},
|
|
{
|
|
"epoch": 0.8073984720546844,
|
|
"grad_norm": 1.1560570001602173,
|
|
"learning_rate": 7.4265098126554065e-06,
|
|
"loss": 0.0489,
|
|
"step": 251
|
|
},
|
|
{
|
|
"epoch": 0.8106151990349819,
|
|
"grad_norm": 1.4469757080078125,
|
|
"learning_rate": 7.401858215313228e-06,
|
|
"loss": 0.0501,
|
|
"step": 252
|
|
},
|
|
{
|
|
"epoch": 0.8138319260152794,
|
|
"grad_norm": 0.8810122609138489,
|
|
"learning_rate": 7.3771304841604764e-06,
|
|
"loss": 0.0317,
|
|
"step": 253
|
|
},
|
|
{
|
|
"epoch": 0.817048652995577,
|
|
"grad_norm": 0.5365167260169983,
|
|
"learning_rate": 7.352327403013779e-06,
|
|
"loss": 0.0414,
|
|
"step": 254
|
|
},
|
|
{
|
|
"epoch": 0.8202653799758746,
|
|
"grad_norm": 0.6032179594039917,
|
|
"learning_rate": 7.327449758078194e-06,
|
|
"loss": 0.0413,
|
|
"step": 255
|
|
},
|
|
{
|
|
"epoch": 0.8234821069561721,
|
|
"grad_norm": 0.7471963763237,
|
|
"learning_rate": 7.302498337922293e-06,
|
|
"loss": 0.0406,
|
|
"step": 256
|
|
},
|
|
{
|
|
"epoch": 0.8266988339364696,
|
|
"grad_norm": 0.5315238237380981,
|
|
"learning_rate": 7.27747393345317e-06,
|
|
"loss": 0.0386,
|
|
"step": 257
|
|
},
|
|
{
|
|
"epoch": 0.8299155609167672,
|
|
"grad_norm": 0.7365145683288574,
|
|
"learning_rate": 7.2523773378913655e-06,
|
|
"loss": 0.0554,
|
|
"step": 258
|
|
},
|
|
{
|
|
"epoch": 0.8331322878970647,
|
|
"grad_norm": 1.0224618911743164,
|
|
"learning_rate": 7.2272093467457226e-06,
|
|
"loss": 0.0434,
|
|
"step": 259
|
|
},
|
|
{
|
|
"epoch": 0.8363490148773622,
|
|
"grad_norm": 1.6006726026535034,
|
|
"learning_rate": 7.201970757788172e-06,
|
|
"loss": 0.0486,
|
|
"step": 260
|
|
},
|
|
{
|
|
"epoch": 0.8395657418576599,
|
|
"grad_norm": 1.2492550611495972,
|
|
"learning_rate": 7.17666237102845e-06,
|
|
"loss": 0.0448,
|
|
"step": 261
|
|
},
|
|
{
|
|
"epoch": 0.8427824688379574,
|
|
"grad_norm": 2.2926084995269775,
|
|
"learning_rate": 7.151284988688731e-06,
|
|
"loss": 0.044,
|
|
"step": 262
|
|
},
|
|
{
|
|
"epoch": 0.8459991958182549,
|
|
"grad_norm": 0.7440332174301147,
|
|
"learning_rate": 7.125839415178204e-06,
|
|
"loss": 0.0477,
|
|
"step": 263
|
|
},
|
|
{
|
|
"epoch": 0.8492159227985525,
|
|
"grad_norm": 0.717848002910614,
|
|
"learning_rate": 7.100326457067576e-06,
|
|
"loss": 0.0396,
|
|
"step": 264
|
|
},
|
|
{
|
|
"epoch": 0.85243264977885,
|
|
"grad_norm": 0.6726682782173157,
|
|
"learning_rate": 7.074746923063497e-06,
|
|
"loss": 0.051,
|
|
"step": 265
|
|
},
|
|
{
|
|
"epoch": 0.8556493767591475,
|
|
"grad_norm": 0.8255169987678528,
|
|
"learning_rate": 7.049101623982938e-06,
|
|
"loss": 0.0445,
|
|
"step": 266
|
|
},
|
|
{
|
|
"epoch": 0.8588661037394452,
|
|
"grad_norm": 0.7894156575202942,
|
|
"learning_rate": 7.02339137272748e-06,
|
|
"loss": 0.0378,
|
|
"step": 267
|
|
},
|
|
{
|
|
"epoch": 0.8620828307197427,
|
|
"grad_norm": 0.8157196640968323,
|
|
"learning_rate": 6.9976169842575526e-06,
|
|
"loss": 0.0395,
|
|
"step": 268
|
|
},
|
|
{
|
|
"epoch": 0.8652995577000402,
|
|
"grad_norm": 0.613278329372406,
|
|
"learning_rate": 6.971779275566593e-06,
|
|
"loss": 0.032,
|
|
"step": 269
|
|
},
|
|
{
|
|
"epoch": 0.8685162846803377,
|
|
"grad_norm": 0.6524770855903625,
|
|
"learning_rate": 6.945879065655164e-06,
|
|
"loss": 0.0463,
|
|
"step": 270
|
|
},
|
|
{
|
|
"epoch": 0.8717330116606353,
|
|
"grad_norm": 0.8012259602546692,
|
|
"learning_rate": 6.919917175504978e-06,
|
|
"loss": 0.0556,
|
|
"step": 271
|
|
},
|
|
{
|
|
"epoch": 0.8749497386409328,
|
|
"grad_norm": 0.8766384720802307,
|
|
"learning_rate": 6.893894428052881e-06,
|
|
"loss": 0.0534,
|
|
"step": 272
|
|
},
|
|
{
|
|
"epoch": 0.8781664656212304,
|
|
"grad_norm": 0.6129264831542969,
|
|
"learning_rate": 6.867811648164769e-06,
|
|
"loss": 0.0486,
|
|
"step": 273
|
|
},
|
|
{
|
|
"epoch": 0.881383192601528,
|
|
"grad_norm": 0.7514522075653076,
|
|
"learning_rate": 6.841669662609437e-06,
|
|
"loss": 0.0464,
|
|
"step": 274
|
|
},
|
|
{
|
|
"epoch": 0.8845999195818255,
|
|
"grad_norm": 0.5673539042472839,
|
|
"learning_rate": 6.815469300032374e-06,
|
|
"loss": 0.0338,
|
|
"step": 275
|
|
},
|
|
{
|
|
"epoch": 0.887816646562123,
|
|
"grad_norm": 0.5022168159484863,
|
|
"learning_rate": 6.789211390929497e-06,
|
|
"loss": 0.0283,
|
|
"step": 276
|
|
},
|
|
{
|
|
"epoch": 0.8910333735424206,
|
|
"grad_norm": 1.195987343788147,
|
|
"learning_rate": 6.762896767620827e-06,
|
|
"loss": 0.0488,
|
|
"step": 277
|
|
},
|
|
{
|
|
"epoch": 0.8942501005227181,
|
|
"grad_norm": 1.0466563701629639,
|
|
"learning_rate": 6.736526264224101e-06,
|
|
"loss": 0.0412,
|
|
"step": 278
|
|
},
|
|
{
|
|
"epoch": 0.8974668275030157,
|
|
"grad_norm": 0.6366952061653137,
|
|
"learning_rate": 6.710100716628345e-06,
|
|
"loss": 0.0403,
|
|
"step": 279
|
|
},
|
|
{
|
|
"epoch": 0.9006835544833133,
|
|
"grad_norm": 1.0278160572052002,
|
|
"learning_rate": 6.6836209624673575e-06,
|
|
"loss": 0.0527,
|
|
"step": 280
|
|
},
|
|
{
|
|
"epoch": 0.9039002814636108,
|
|
"grad_norm": 0.8810920715332031,
|
|
"learning_rate": 6.657087841093179e-06,
|
|
"loss": 0.0517,
|
|
"step": 281
|
|
},
|
|
{
|
|
"epoch": 0.9071170084439083,
|
|
"grad_norm": 1.306876301765442,
|
|
"learning_rate": 6.6305021935494755e-06,
|
|
"loss": 0.041,
|
|
"step": 282
|
|
},
|
|
{
|
|
"epoch": 0.9103337354242059,
|
|
"grad_norm": 1.134096622467041,
|
|
"learning_rate": 6.603864862544879e-06,
|
|
"loss": 0.0474,
|
|
"step": 283
|
|
},
|
|
{
|
|
"epoch": 0.9135504624045034,
|
|
"grad_norm": 1.364350438117981,
|
|
"learning_rate": 6.5771766924262795e-06,
|
|
"loss": 0.0418,
|
|
"step": 284
|
|
},
|
|
{
|
|
"epoch": 0.916767189384801,
|
|
"grad_norm": 1.0049413442611694,
|
|
"learning_rate": 6.5504385291520554e-06,
|
|
"loss": 0.0386,
|
|
"step": 285
|
|
},
|
|
{
|
|
"epoch": 0.9199839163650985,
|
|
"grad_norm": 0.5477365255355835,
|
|
"learning_rate": 6.523651220265269e-06,
|
|
"loss": 0.0394,
|
|
"step": 286
|
|
},
|
|
{
|
|
"epoch": 0.9232006433453961,
|
|
"grad_norm": 1.8029471635818481,
|
|
"learning_rate": 6.496815614866792e-06,
|
|
"loss": 0.0486,
|
|
"step": 287
|
|
},
|
|
{
|
|
"epoch": 0.9264173703256936,
|
|
"grad_norm": 1.1982808113098145,
|
|
"learning_rate": 6.469932563588386e-06,
|
|
"loss": 0.0385,
|
|
"step": 288
|
|
},
|
|
{
|
|
"epoch": 0.9296340973059911,
|
|
"grad_norm": 1.8659429550170898,
|
|
"learning_rate": 6.443002918565754e-06,
|
|
"loss": 0.05,
|
|
"step": 289
|
|
},
|
|
{
|
|
"epoch": 0.9328508242862887,
|
|
"grad_norm": 0.9505304098129272,
|
|
"learning_rate": 6.41602753341152e-06,
|
|
"loss": 0.0439,
|
|
"step": 290
|
|
},
|
|
{
|
|
"epoch": 0.9360675512665863,
|
|
"grad_norm": 1.1647857427597046,
|
|
"learning_rate": 6.389007263188176e-06,
|
|
"loss": 0.0472,
|
|
"step": 291
|
|
},
|
|
{
|
|
"epoch": 0.9392842782468838,
|
|
"grad_norm": 1.5960369110107422,
|
|
"learning_rate": 6.361942964380967e-06,
|
|
"loss": 0.0454,
|
|
"step": 292
|
|
},
|
|
{
|
|
"epoch": 0.9425010052271814,
|
|
"grad_norm": 0.6902590990066528,
|
|
"learning_rate": 6.334835494870759e-06,
|
|
"loss": 0.0351,
|
|
"step": 293
|
|
},
|
|
{
|
|
"epoch": 0.9457177322074789,
|
|
"grad_norm": 0.8172485828399658,
|
|
"learning_rate": 6.307685713906835e-06,
|
|
"loss": 0.0344,
|
|
"step": 294
|
|
},
|
|
{
|
|
"epoch": 0.9489344591877764,
|
|
"grad_norm": 0.46139606833457947,
|
|
"learning_rate": 6.2804944820796596e-06,
|
|
"loss": 0.0308,
|
|
"step": 295
|
|
},
|
|
{
|
|
"epoch": 0.952151186168074,
|
|
"grad_norm": 0.7500118613243103,
|
|
"learning_rate": 6.2532626612936035e-06,
|
|
"loss": 0.0541,
|
|
"step": 296
|
|
},
|
|
{
|
|
"epoch": 0.9553679131483716,
|
|
"grad_norm": 0.693679928779602,
|
|
"learning_rate": 6.225991114739622e-06,
|
|
"loss": 0.0351,
|
|
"step": 297
|
|
},
|
|
{
|
|
"epoch": 0.9585846401286691,
|
|
"grad_norm": 1.401061773300171,
|
|
"learning_rate": 6.1986807068678926e-06,
|
|
"loss": 0.0458,
|
|
"step": 298
|
|
},
|
|
{
|
|
"epoch": 0.9618013671089666,
|
|
"grad_norm": 0.5701402425765991,
|
|
"learning_rate": 6.171332303360411e-06,
|
|
"loss": 0.0349,
|
|
"step": 299
|
|
},
|
|
{
|
|
"epoch": 0.9650180940892642,
|
|
"grad_norm": 1.6502150297164917,
|
|
"learning_rate": 6.143946771103561e-06,
|
|
"loss": 0.0553,
|
|
"step": 300
|
|
},
|
|
{
|
|
"epoch": 0.9682348210695617,
|
|
"grad_norm": 0.8593463897705078,
|
|
"learning_rate": 6.11652497816062e-06,
|
|
"loss": 0.0372,
|
|
"step": 301
|
|
},
|
|
{
|
|
"epoch": 0.9714515480498592,
|
|
"grad_norm": 0.8635258674621582,
|
|
"learning_rate": 6.089067793744258e-06,
|
|
"loss": 0.0386,
|
|
"step": 302
|
|
},
|
|
{
|
|
"epoch": 0.9746682750301568,
|
|
"grad_norm": 1.058119535446167,
|
|
"learning_rate": 6.061576088188981e-06,
|
|
"loss": 0.0407,
|
|
"step": 303
|
|
},
|
|
{
|
|
"epoch": 0.9778850020104544,
|
|
"grad_norm": 1.316687822341919,
|
|
"learning_rate": 6.034050732923538e-06,
|
|
"loss": 0.0405,
|
|
"step": 304
|
|
},
|
|
{
|
|
"epoch": 0.9811017289907519,
|
|
"grad_norm": 0.6700626611709595,
|
|
"learning_rate": 6.006492600443301e-06,
|
|
"loss": 0.0456,
|
|
"step": 305
|
|
},
|
|
{
|
|
"epoch": 0.9843184559710495,
|
|
"grad_norm": 1.0319327116012573,
|
|
"learning_rate": 5.978902564282616e-06,
|
|
"loss": 0.0476,
|
|
"step": 306
|
|
},
|
|
{
|
|
"epoch": 0.987535182951347,
|
|
"grad_norm": 0.5660614371299744,
|
|
"learning_rate": 5.951281498987106e-06,
|
|
"loss": 0.0431,
|
|
"step": 307
|
|
},
|
|
{
|
|
"epoch": 0.9907519099316445,
|
|
"grad_norm": 0.5249197483062744,
|
|
"learning_rate": 5.923630280085948e-06,
|
|
"loss": 0.0416,
|
|
"step": 308
|
|
},
|
|
{
|
|
"epoch": 0.9939686369119421,
|
|
"grad_norm": 0.5636727213859558,
|
|
"learning_rate": 5.895949784064126e-06,
|
|
"loss": 0.0316,
|
|
"step": 309
|
|
},
|
|
{
|
|
"epoch": 0.9971853638922397,
|
|
"grad_norm": 0.5312305688858032,
|
|
"learning_rate": 5.8682408883346535e-06,
|
|
"loss": 0.0393,
|
|
"step": 310
|
|
},
|
|
{
|
|
"epoch": 1.0,
|
|
"grad_norm": 0.8194127678871155,
|
|
"learning_rate": 5.840504471210742e-06,
|
|
"loss": 0.0507,
|
|
"step": 311
|
|
},
|
|
{
|
|
"epoch": 1.0032167269802976,
|
|
"grad_norm": 0.4939586818218231,
|
|
"learning_rate": 5.8127414118779825e-06,
|
|
"loss": 0.0331,
|
|
"step": 312
|
|
},
|
|
{
|
|
"epoch": 1.006433453960595,
|
|
"grad_norm": 0.9499027729034424,
|
|
"learning_rate": 5.7849525903664636e-06,
|
|
"loss": 0.0234,
|
|
"step": 313
|
|
},
|
|
{
|
|
"epoch": 1.0096501809408926,
|
|
"grad_norm": 1.0846751928329468,
|
|
"learning_rate": 5.757138887522884e-06,
|
|
"loss": 0.0244,
|
|
"step": 314
|
|
},
|
|
{
|
|
"epoch": 1.0128669079211903,
|
|
"grad_norm": 0.39767810702323914,
|
|
"learning_rate": 5.729301184982622e-06,
|
|
"loss": 0.0334,
|
|
"step": 315
|
|
},
|
|
{
|
|
"epoch": 1.0160836349014877,
|
|
"grad_norm": 0.38344645500183105,
|
|
"learning_rate": 5.701440365141799e-06,
|
|
"loss": 0.0301,
|
|
"step": 316
|
|
},
|
|
{
|
|
"epoch": 1.0193003618817853,
|
|
"grad_norm": 0.7626691460609436,
|
|
"learning_rate": 5.673557311129306e-06,
|
|
"loss": 0.0322,
|
|
"step": 317
|
|
},
|
|
{
|
|
"epoch": 1.022517088862083,
|
|
"grad_norm": 0.5533730387687683,
|
|
"learning_rate": 5.645652906778808e-06,
|
|
"loss": 0.0304,
|
|
"step": 318
|
|
},
|
|
{
|
|
"epoch": 1.0257338158423803,
|
|
"grad_norm": 0.7238726615905762,
|
|
"learning_rate": 5.617728036600734e-06,
|
|
"loss": 0.0219,
|
|
"step": 319
|
|
},
|
|
{
|
|
"epoch": 1.028950542822678,
|
|
"grad_norm": 0.6678081750869751,
|
|
"learning_rate": 5.5897835857542315e-06,
|
|
"loss": 0.0271,
|
|
"step": 320
|
|
},
|
|
{
|
|
"epoch": 1.0321672698029756,
|
|
"grad_norm": 1.125412940979004,
|
|
"learning_rate": 5.561820440019117e-06,
|
|
"loss": 0.0281,
|
|
"step": 321
|
|
},
|
|
{
|
|
"epoch": 1.035383996783273,
|
|
"grad_norm": 0.9026764631271362,
|
|
"learning_rate": 5.533839485767795e-06,
|
|
"loss": 0.0181,
|
|
"step": 322
|
|
},
|
|
{
|
|
"epoch": 1.0386007237635706,
|
|
"grad_norm": 0.6368532776832581,
|
|
"learning_rate": 5.505841609937162e-06,
|
|
"loss": 0.0261,
|
|
"step": 323
|
|
},
|
|
{
|
|
"epoch": 1.041817450743868,
|
|
"grad_norm": 0.48171374201774597,
|
|
"learning_rate": 5.477827700000492e-06,
|
|
"loss": 0.0168,
|
|
"step": 324
|
|
},
|
|
{
|
|
"epoch": 1.0450341777241656,
|
|
"grad_norm": 0.7617553472518921,
|
|
"learning_rate": 5.449798643939305e-06,
|
|
"loss": 0.0294,
|
|
"step": 325
|
|
},
|
|
{
|
|
"epoch": 1.0482509047044632,
|
|
"grad_norm": 0.7659939527511597,
|
|
"learning_rate": 5.421755330215223e-06,
|
|
"loss": 0.0233,
|
|
"step": 326
|
|
},
|
|
{
|
|
"epoch": 1.0514676316847607,
|
|
"grad_norm": 1.1631779670715332,
|
|
"learning_rate": 5.39369864774181e-06,
|
|
"loss": 0.031,
|
|
"step": 327
|
|
},
|
|
{
|
|
"epoch": 1.0546843586650583,
|
|
"grad_norm": 0.5692664384841919,
|
|
"learning_rate": 5.365629485856381e-06,
|
|
"loss": 0.017,
|
|
"step": 328
|
|
},
|
|
{
|
|
"epoch": 1.057901085645356,
|
|
"grad_norm": 0.8848438858985901,
|
|
"learning_rate": 5.337548734291827e-06,
|
|
"loss": 0.0372,
|
|
"step": 329
|
|
},
|
|
{
|
|
"epoch": 1.0611178126256533,
|
|
"grad_norm": 0.6400728225708008,
|
|
"learning_rate": 5.30945728314841e-06,
|
|
"loss": 0.0213,
|
|
"step": 330
|
|
},
|
|
{
|
|
"epoch": 1.064334539605951,
|
|
"grad_norm": 0.9902735352516174,
|
|
"learning_rate": 5.281356022865542e-06,
|
|
"loss": 0.0262,
|
|
"step": 331
|
|
},
|
|
{
|
|
"epoch": 1.0675512665862485,
|
|
"grad_norm": 0.5511563420295715,
|
|
"learning_rate": 5.253245844193564e-06,
|
|
"loss": 0.0212,
|
|
"step": 332
|
|
},
|
|
{
|
|
"epoch": 1.070767993566546,
|
|
"grad_norm": 0.7277644276618958,
|
|
"learning_rate": 5.225127638165514e-06,
|
|
"loss": 0.0164,
|
|
"step": 333
|
|
},
|
|
{
|
|
"epoch": 1.0739847205468436,
|
|
"grad_norm": 0.5540727376937866,
|
|
"learning_rate": 5.197002296068878e-06,
|
|
"loss": 0.024,
|
|
"step": 334
|
|
},
|
|
{
|
|
"epoch": 1.0772014475271412,
|
|
"grad_norm": 0.55019611120224,
|
|
"learning_rate": 5.168870709417342e-06,
|
|
"loss": 0.0203,
|
|
"step": 335
|
|
},
|
|
{
|
|
"epoch": 1.0804181745074386,
|
|
"grad_norm": 0.5772176384925842,
|
|
"learning_rate": 5.140733769922525e-06,
|
|
"loss": 0.021,
|
|
"step": 336
|
|
},
|
|
{
|
|
"epoch": 1.0836349014877362,
|
|
"grad_norm": 0.7441113591194153,
|
|
"learning_rate": 5.112592369465731e-06,
|
|
"loss": 0.0372,
|
|
"step": 337
|
|
},
|
|
{
|
|
"epoch": 1.0868516284680338,
|
|
"grad_norm": 0.5592713952064514,
|
|
"learning_rate": 5.084447400069656e-06,
|
|
"loss": 0.02,
|
|
"step": 338
|
|
},
|
|
{
|
|
"epoch": 1.0900683554483312,
|
|
"grad_norm": 0.6708924770355225,
|
|
"learning_rate": 5.0562997538701295e-06,
|
|
"loss": 0.0207,
|
|
"step": 339
|
|
},
|
|
{
|
|
"epoch": 1.0932850824286289,
|
|
"grad_norm": 0.6388556361198425,
|
|
"learning_rate": 5.0281503230878304e-06,
|
|
"loss": 0.0234,
|
|
"step": 340
|
|
},
|
|
{
|
|
"epoch": 1.0965018094089265,
|
|
"grad_norm": 0.5663427114486694,
|
|
"learning_rate": 5e-06,
|
|
"loss": 0.02,
|
|
"step": 341
|
|
},
|
|
{
|
|
"epoch": 1.099718536389224,
|
|
"grad_norm": 1.1264530420303345,
|
|
"learning_rate": 4.971849676912172e-06,
|
|
"loss": 0.0243,
|
|
"step": 342
|
|
},
|
|
{
|
|
"epoch": 1.1029352633695215,
|
|
"grad_norm": 1.2352030277252197,
|
|
"learning_rate": 4.943700246129871e-06,
|
|
"loss": 0.0297,
|
|
"step": 343
|
|
},
|
|
{
|
|
"epoch": 1.1061519903498191,
|
|
"grad_norm": 1.2657285928726196,
|
|
"learning_rate": 4.915552599930345e-06,
|
|
"loss": 0.0204,
|
|
"step": 344
|
|
},
|
|
{
|
|
"epoch": 1.1093687173301165,
|
|
"grad_norm": 0.7982221245765686,
|
|
"learning_rate": 4.887407630534271e-06,
|
|
"loss": 0.0306,
|
|
"step": 345
|
|
},
|
|
{
|
|
"epoch": 1.1125854443104142,
|
|
"grad_norm": 0.7376164197921753,
|
|
"learning_rate": 4.859266230077474e-06,
|
|
"loss": 0.0299,
|
|
"step": 346
|
|
},
|
|
{
|
|
"epoch": 1.1158021712907118,
|
|
"grad_norm": 0.7835693359375,
|
|
"learning_rate": 4.83112929058266e-06,
|
|
"loss": 0.0231,
|
|
"step": 347
|
|
},
|
|
{
|
|
"epoch": 1.1190188982710092,
|
|
"grad_norm": 1.1944645643234253,
|
|
"learning_rate": 4.802997703931124e-06,
|
|
"loss": 0.0277,
|
|
"step": 348
|
|
},
|
|
{
|
|
"epoch": 1.1222356252513068,
|
|
"grad_norm": 0.6810235977172852,
|
|
"learning_rate": 4.7748723618344865e-06,
|
|
"loss": 0.0261,
|
|
"step": 349
|
|
},
|
|
{
|
|
"epoch": 1.1254523522316044,
|
|
"grad_norm": 0.6911215782165527,
|
|
"learning_rate": 4.746754155806437e-06,
|
|
"loss": 0.0213,
|
|
"step": 350
|
|
},
|
|
{
|
|
"epoch": 1.1286690792119018,
|
|
"grad_norm": 0.8558758497238159,
|
|
"learning_rate": 4.71864397713446e-06,
|
|
"loss": 0.0238,
|
|
"step": 351
|
|
},
|
|
{
|
|
"epoch": 1.1318858061921995,
|
|
"grad_norm": 0.9453890919685364,
|
|
"learning_rate": 4.6905427168515914e-06,
|
|
"loss": 0.0303,
|
|
"step": 352
|
|
},
|
|
{
|
|
"epoch": 1.135102533172497,
|
|
"grad_norm": 0.666305661201477,
|
|
"learning_rate": 4.662451265708174e-06,
|
|
"loss": 0.0188,
|
|
"step": 353
|
|
},
|
|
{
|
|
"epoch": 1.1383192601527945,
|
|
"grad_norm": 0.7106337547302246,
|
|
"learning_rate": 4.63437051414362e-06,
|
|
"loss": 0.0305,
|
|
"step": 354
|
|
},
|
|
{
|
|
"epoch": 1.1415359871330921,
|
|
"grad_norm": 0.5249149203300476,
|
|
"learning_rate": 4.606301352258192e-06,
|
|
"loss": 0.0186,
|
|
"step": 355
|
|
},
|
|
{
|
|
"epoch": 1.1447527141133897,
|
|
"grad_norm": 0.6891714334487915,
|
|
"learning_rate": 4.5782446697847775e-06,
|
|
"loss": 0.0193,
|
|
"step": 356
|
|
},
|
|
{
|
|
"epoch": 1.1479694410936871,
|
|
"grad_norm": 1.2555615901947021,
|
|
"learning_rate": 4.550201356060695e-06,
|
|
"loss": 0.0274,
|
|
"step": 357
|
|
},
|
|
{
|
|
"epoch": 1.1511861680739848,
|
|
"grad_norm": 1.113802433013916,
|
|
"learning_rate": 4.52217229999951e-06,
|
|
"loss": 0.0256,
|
|
"step": 358
|
|
},
|
|
{
|
|
"epoch": 1.1544028950542822,
|
|
"grad_norm": 0.6190158724784851,
|
|
"learning_rate": 4.49415839006284e-06,
|
|
"loss": 0.0265,
|
|
"step": 359
|
|
},
|
|
{
|
|
"epoch": 1.1576196220345798,
|
|
"grad_norm": 0.8673491477966309,
|
|
"learning_rate": 4.466160514232206e-06,
|
|
"loss": 0.0341,
|
|
"step": 360
|
|
},
|
|
{
|
|
"epoch": 1.1608363490148774,
|
|
"grad_norm": 0.4235200583934784,
|
|
"learning_rate": 4.438179559980885e-06,
|
|
"loss": 0.0163,
|
|
"step": 361
|
|
},
|
|
{
|
|
"epoch": 1.1640530759951748,
|
|
"grad_norm": 0.9669196605682373,
|
|
"learning_rate": 4.410216414245771e-06,
|
|
"loss": 0.0357,
|
|
"step": 362
|
|
},
|
|
{
|
|
"epoch": 1.1672698029754724,
|
|
"grad_norm": 0.8336294293403625,
|
|
"learning_rate": 4.382271963399268e-06,
|
|
"loss": 0.0161,
|
|
"step": 363
|
|
},
|
|
{
|
|
"epoch": 1.17048652995577,
|
|
"grad_norm": 0.8520212769508362,
|
|
"learning_rate": 4.354347093221194e-06,
|
|
"loss": 0.0206,
|
|
"step": 364
|
|
},
|
|
{
|
|
"epoch": 1.1737032569360675,
|
|
"grad_norm": 0.9719181060791016,
|
|
"learning_rate": 4.326442688870697e-06,
|
|
"loss": 0.0207,
|
|
"step": 365
|
|
},
|
|
{
|
|
"epoch": 1.176919983916365,
|
|
"grad_norm": 0.6609673500061035,
|
|
"learning_rate": 4.298559634858202e-06,
|
|
"loss": 0.0288,
|
|
"step": 366
|
|
},
|
|
{
|
|
"epoch": 1.1801367108966627,
|
|
"grad_norm": 0.9857617616653442,
|
|
"learning_rate": 4.270698815017379e-06,
|
|
"loss": 0.0279,
|
|
"step": 367
|
|
},
|
|
{
|
|
"epoch": 1.1833534378769601,
|
|
"grad_norm": 0.788272500038147,
|
|
"learning_rate": 4.2428611124771184e-06,
|
|
"loss": 0.0144,
|
|
"step": 368
|
|
},
|
|
{
|
|
"epoch": 1.1865701648572577,
|
|
"grad_norm": 0.8060084581375122,
|
|
"learning_rate": 4.2150474096335356e-06,
|
|
"loss": 0.023,
|
|
"step": 369
|
|
},
|
|
{
|
|
"epoch": 1.1897868918375554,
|
|
"grad_norm": 1.2030549049377441,
|
|
"learning_rate": 4.187258588122019e-06,
|
|
"loss": 0.0355,
|
|
"step": 370
|
|
},
|
|
{
|
|
"epoch": 1.1930036188178528,
|
|
"grad_norm": 0.6166083812713623,
|
|
"learning_rate": 4.15949552878926e-06,
|
|
"loss": 0.0165,
|
|
"step": 371
|
|
},
|
|
{
|
|
"epoch": 1.1962203457981504,
|
|
"grad_norm": 1.5844203233718872,
|
|
"learning_rate": 4.131759111665349e-06,
|
|
"loss": 0.0323,
|
|
"step": 372
|
|
},
|
|
{
|
|
"epoch": 1.199437072778448,
|
|
"grad_norm": 0.9516481757164001,
|
|
"learning_rate": 4.104050215935875e-06,
|
|
"loss": 0.0306,
|
|
"step": 373
|
|
},
|
|
{
|
|
"epoch": 1.2026537997587454,
|
|
"grad_norm": 1.0186715126037598,
|
|
"learning_rate": 4.076369719914055e-06,
|
|
"loss": 0.0252,
|
|
"step": 374
|
|
},
|
|
{
|
|
"epoch": 1.205870526739043,
|
|
"grad_norm": 0.891091525554657,
|
|
"learning_rate": 4.048718501012895e-06,
|
|
"loss": 0.0168,
|
|
"step": 375
|
|
},
|
|
{
|
|
"epoch": 1.2090872537193404,
|
|
"grad_norm": 0.6966618299484253,
|
|
"learning_rate": 4.021097435717386e-06,
|
|
"loss": 0.0297,
|
|
"step": 376
|
|
},
|
|
{
|
|
"epoch": 1.212303980699638,
|
|
"grad_norm": 0.9956583380699158,
|
|
"learning_rate": 3.993507399556699e-06,
|
|
"loss": 0.0189,
|
|
"step": 377
|
|
},
|
|
{
|
|
"epoch": 1.2155207076799357,
|
|
"grad_norm": 0.6981390714645386,
|
|
"learning_rate": 3.965949267076465e-06,
|
|
"loss": 0.0175,
|
|
"step": 378
|
|
},
|
|
{
|
|
"epoch": 1.218737434660233,
|
|
"grad_norm": 0.9271097779273987,
|
|
"learning_rate": 3.938423911811021e-06,
|
|
"loss": 0.0215,
|
|
"step": 379
|
|
},
|
|
{
|
|
"epoch": 1.2219541616405307,
|
|
"grad_norm": 1.0739299058914185,
|
|
"learning_rate": 3.910932206255742e-06,
|
|
"loss": 0.0304,
|
|
"step": 380
|
|
},
|
|
{
|
|
"epoch": 1.2251708886208283,
|
|
"grad_norm": 0.7238112092018127,
|
|
"learning_rate": 3.883475021839382e-06,
|
|
"loss": 0.0269,
|
|
"step": 381
|
|
},
|
|
{
|
|
"epoch": 1.2283876156011257,
|
|
"grad_norm": 0.8449429273605347,
|
|
"learning_rate": 3.856053228896442e-06,
|
|
"loss": 0.0253,
|
|
"step": 382
|
|
},
|
|
{
|
|
"epoch": 1.2316043425814234,
|
|
"grad_norm": 1.0511232614517212,
|
|
"learning_rate": 3.8286676966395895e-06,
|
|
"loss": 0.0286,
|
|
"step": 383
|
|
},
|
|
{
|
|
"epoch": 1.234821069561721,
|
|
"grad_norm": 0.6601850390434265,
|
|
"learning_rate": 3.8013192931321095e-06,
|
|
"loss": 0.0161,
|
|
"step": 384
|
|
},
|
|
{
|
|
"epoch": 1.2380377965420184,
|
|
"grad_norm": 1.7054966688156128,
|
|
"learning_rate": 3.77400888526038e-06,
|
|
"loss": 0.0382,
|
|
"step": 385
|
|
},
|
|
{
|
|
"epoch": 1.241254523522316,
|
|
"grad_norm": 0.5929433703422546,
|
|
"learning_rate": 3.7467373387063973e-06,
|
|
"loss": 0.027,
|
|
"step": 386
|
|
},
|
|
{
|
|
"epoch": 1.2444712505026136,
|
|
"grad_norm": 0.6725859642028809,
|
|
"learning_rate": 3.719505517920342e-06,
|
|
"loss": 0.0283,
|
|
"step": 387
|
|
},
|
|
{
|
|
"epoch": 1.247687977482911,
|
|
"grad_norm": 0.6051467657089233,
|
|
"learning_rate": 3.692314286093167e-06,
|
|
"loss": 0.024,
|
|
"step": 388
|
|
},
|
|
{
|
|
"epoch": 1.2509047044632087,
|
|
"grad_norm": 0.7530882358551025,
|
|
"learning_rate": 3.6651645051292415e-06,
|
|
"loss": 0.0226,
|
|
"step": 389
|
|
},
|
|
{
|
|
"epoch": 1.2541214314435063,
|
|
"grad_norm": 0.6657713651657104,
|
|
"learning_rate": 3.6380570356190346e-06,
|
|
"loss": 0.0197,
|
|
"step": 390
|
|
},
|
|
{
|
|
"epoch": 1.2573381584238037,
|
|
"grad_norm": 1.1519615650177002,
|
|
"learning_rate": 3.610992736811827e-06,
|
|
"loss": 0.0188,
|
|
"step": 391
|
|
},
|
|
{
|
|
"epoch": 1.2605548854041013,
|
|
"grad_norm": 0.8955692648887634,
|
|
"learning_rate": 3.58397246658848e-06,
|
|
"loss": 0.0193,
|
|
"step": 392
|
|
},
|
|
{
|
|
"epoch": 1.263771612384399,
|
|
"grad_norm": 1.1424133777618408,
|
|
"learning_rate": 3.556997081434248e-06,
|
|
"loss": 0.025,
|
|
"step": 393
|
|
},
|
|
{
|
|
"epoch": 1.2669883393646963,
|
|
"grad_norm": 0.634136974811554,
|
|
"learning_rate": 3.5300674364116173e-06,
|
|
"loss": 0.0237,
|
|
"step": 394
|
|
},
|
|
{
|
|
"epoch": 1.270205066344994,
|
|
"grad_norm": 0.9286861419677734,
|
|
"learning_rate": 3.5031843851332105e-06,
|
|
"loss": 0.0231,
|
|
"step": 395
|
|
},
|
|
{
|
|
"epoch": 1.2734217933252916,
|
|
"grad_norm": 0.7624120712280273,
|
|
"learning_rate": 3.476348779734732e-06,
|
|
"loss": 0.0246,
|
|
"step": 396
|
|
},
|
|
{
|
|
"epoch": 1.276638520305589,
|
|
"grad_norm": 0.8140252828598022,
|
|
"learning_rate": 3.449561470847947e-06,
|
|
"loss": 0.0262,
|
|
"step": 397
|
|
},
|
|
{
|
|
"epoch": 1.2798552472858866,
|
|
"grad_norm": 1.130080223083496,
|
|
"learning_rate": 3.4228233075737225e-06,
|
|
"loss": 0.0294,
|
|
"step": 398
|
|
},
|
|
{
|
|
"epoch": 1.2830719742661842,
|
|
"grad_norm": 0.8101189136505127,
|
|
"learning_rate": 3.3961351374551234e-06,
|
|
"loss": 0.0279,
|
|
"step": 399
|
|
},
|
|
{
|
|
"epoch": 1.2862887012464816,
|
|
"grad_norm": 0.5413293838500977,
|
|
"learning_rate": 3.3694978064505258e-06,
|
|
"loss": 0.0104,
|
|
"step": 400
|
|
},
|
|
{
|
|
"epoch": 1.2895054282267793,
|
|
"grad_norm": 0.940980851650238,
|
|
"learning_rate": 3.3429121589068213e-06,
|
|
"loss": 0.023,
|
|
"step": 401
|
|
},
|
|
{
|
|
"epoch": 1.292722155207077,
|
|
"grad_norm": 1.4094127416610718,
|
|
"learning_rate": 3.316379037532644e-06,
|
|
"loss": 0.0238,
|
|
"step": 402
|
|
},
|
|
{
|
|
"epoch": 1.2959388821873743,
|
|
"grad_norm": 0.6321690082550049,
|
|
"learning_rate": 3.289899283371657e-06,
|
|
"loss": 0.0233,
|
|
"step": 403
|
|
},
|
|
{
|
|
"epoch": 1.299155609167672,
|
|
"grad_norm": 0.7912178635597229,
|
|
"learning_rate": 3.2634737357758994e-06,
|
|
"loss": 0.0247,
|
|
"step": 404
|
|
},
|
|
{
|
|
"epoch": 1.3023723361479695,
|
|
"grad_norm": 0.808509111404419,
|
|
"learning_rate": 3.2371032323791757e-06,
|
|
"loss": 0.0209,
|
|
"step": 405
|
|
},
|
|
{
|
|
"epoch": 1.305589063128267,
|
|
"grad_norm": 0.5239757895469666,
|
|
"learning_rate": 3.2107886090705035e-06,
|
|
"loss": 0.0188,
|
|
"step": 406
|
|
},
|
|
{
|
|
"epoch": 1.3088057901085646,
|
|
"grad_norm": 0.7358213067054749,
|
|
"learning_rate": 3.1845306999676274e-06,
|
|
"loss": 0.0328,
|
|
"step": 407
|
|
},
|
|
{
|
|
"epoch": 1.3120225170888622,
|
|
"grad_norm": 0.8183391690254211,
|
|
"learning_rate": 3.158330337390565e-06,
|
|
"loss": 0.0263,
|
|
"step": 408
|
|
},
|
|
{
|
|
"epoch": 1.3152392440691596,
|
|
"grad_norm": 0.7185299396514893,
|
|
"learning_rate": 3.132188351835232e-06,
|
|
"loss": 0.0316,
|
|
"step": 409
|
|
},
|
|
{
|
|
"epoch": 1.3184559710494572,
|
|
"grad_norm": 0.7911690473556519,
|
|
"learning_rate": 3.10610557194712e-06,
|
|
"loss": 0.0229,
|
|
"step": 410
|
|
},
|
|
{
|
|
"epoch": 1.3216726980297548,
|
|
"grad_norm": 0.5570195317268372,
|
|
"learning_rate": 3.080082824495024e-06,
|
|
"loss": 0.0276,
|
|
"step": 411
|
|
},
|
|
{
|
|
"epoch": 1.3248894250100522,
|
|
"grad_norm": 0.760032057762146,
|
|
"learning_rate": 3.0541209343448373e-06,
|
|
"loss": 0.0228,
|
|
"step": 412
|
|
},
|
|
{
|
|
"epoch": 1.3281061519903499,
|
|
"grad_norm": 0.6067129373550415,
|
|
"learning_rate": 3.0282207244334084e-06,
|
|
"loss": 0.0211,
|
|
"step": 413
|
|
},
|
|
{
|
|
"epoch": 1.3313228789706475,
|
|
"grad_norm": 0.8982436656951904,
|
|
"learning_rate": 3.0023830157424504e-06,
|
|
"loss": 0.0269,
|
|
"step": 414
|
|
},
|
|
{
|
|
"epoch": 1.334539605950945,
|
|
"grad_norm": 0.9359907507896423,
|
|
"learning_rate": 2.97660862727252e-06,
|
|
"loss": 0.0226,
|
|
"step": 415
|
|
},
|
|
{
|
|
"epoch": 1.3377563329312425,
|
|
"grad_norm": 0.5233786106109619,
|
|
"learning_rate": 2.950898376017064e-06,
|
|
"loss": 0.0226,
|
|
"step": 416
|
|
},
|
|
{
|
|
"epoch": 1.3409730599115401,
|
|
"grad_norm": 0.5477368235588074,
|
|
"learning_rate": 2.9252530769365053e-06,
|
|
"loss": 0.0177,
|
|
"step": 417
|
|
},
|
|
{
|
|
"epoch": 1.3441897868918375,
|
|
"grad_norm": 0.7857730984687805,
|
|
"learning_rate": 2.8996735429324256e-06,
|
|
"loss": 0.0208,
|
|
"step": 418
|
|
},
|
|
{
|
|
"epoch": 1.3474065138721352,
|
|
"grad_norm": 0.6837164163589478,
|
|
"learning_rate": 2.874160584821798e-06,
|
|
"loss": 0.0204,
|
|
"step": 419
|
|
},
|
|
{
|
|
"epoch": 1.3506232408524326,
|
|
"grad_norm": 0.5833801627159119,
|
|
"learning_rate": 2.848715011311271e-06,
|
|
"loss": 0.0179,
|
|
"step": 420
|
|
},
|
|
{
|
|
"epoch": 1.3538399678327302,
|
|
"grad_norm": 0.7275317311286926,
|
|
"learning_rate": 2.823337628971551e-06,
|
|
"loss": 0.0176,
|
|
"step": 421
|
|
},
|
|
{
|
|
"epoch": 1.3570566948130278,
|
|
"grad_norm": 1.1249991655349731,
|
|
"learning_rate": 2.7980292422118282e-06,
|
|
"loss": 0.0264,
|
|
"step": 422
|
|
},
|
|
{
|
|
"epoch": 1.3602734217933252,
|
|
"grad_norm": 0.8892255425453186,
|
|
"learning_rate": 2.7727906532542783e-06,
|
|
"loss": 0.021,
|
|
"step": 423
|
|
},
|
|
{
|
|
"epoch": 1.3634901487736228,
|
|
"grad_norm": 0.8986218571662903,
|
|
"learning_rate": 2.7476226621086354e-06,
|
|
"loss": 0.0209,
|
|
"step": 424
|
|
},
|
|
{
|
|
"epoch": 1.3667068757539202,
|
|
"grad_norm": 0.8153395056724548,
|
|
"learning_rate": 2.72252606654683e-06,
|
|
"loss": 0.0209,
|
|
"step": 425
|
|
},
|
|
{
|
|
"epoch": 1.3699236027342179,
|
|
"grad_norm": 0.8225369453430176,
|
|
"learning_rate": 2.697501662077707e-06,
|
|
"loss": 0.0299,
|
|
"step": 426
|
|
},
|
|
{
|
|
"epoch": 1.3731403297145155,
|
|
"grad_norm": 0.8588507175445557,
|
|
"learning_rate": 2.6725502419218084e-06,
|
|
"loss": 0.0346,
|
|
"step": 427
|
|
},
|
|
{
|
|
"epoch": 1.376357056694813,
|
|
"grad_norm": 1.9007744789123535,
|
|
"learning_rate": 2.6476725969862227e-06,
|
|
"loss": 0.0418,
|
|
"step": 428
|
|
},
|
|
{
|
|
"epoch": 1.3795737836751105,
|
|
"grad_norm": 1.121382474899292,
|
|
"learning_rate": 2.622869515839524e-06,
|
|
"loss": 0.0179,
|
|
"step": 429
|
|
},
|
|
{
|
|
"epoch": 1.3827905106554081,
|
|
"grad_norm": 0.7519135475158691,
|
|
"learning_rate": 2.5981417846867753e-06,
|
|
"loss": 0.0242,
|
|
"step": 430
|
|
},
|
|
{
|
|
"epoch": 1.3860072376357055,
|
|
"grad_norm": 0.7129932641983032,
|
|
"learning_rate": 2.573490187344596e-06,
|
|
"loss": 0.0137,
|
|
"step": 431
|
|
},
|
|
{
|
|
"epoch": 1.3892239646160032,
|
|
"grad_norm": 0.7935872673988342,
|
|
"learning_rate": 2.548915505216333e-06,
|
|
"loss": 0.0287,
|
|
"step": 432
|
|
},
|
|
{
|
|
"epoch": 1.3924406915963008,
|
|
"grad_norm": 1.06389582157135,
|
|
"learning_rate": 2.524418517267283e-06,
|
|
"loss": 0.0251,
|
|
"step": 433
|
|
},
|
|
{
|
|
"epoch": 1.3956574185765982,
|
|
"grad_norm": 0.8240230679512024,
|
|
"learning_rate": 2.5000000000000015e-06,
|
|
"loss": 0.0297,
|
|
"step": 434
|
|
},
|
|
{
|
|
"epoch": 1.3988741455568958,
|
|
"grad_norm": 0.9097860455513,
|
|
"learning_rate": 2.4756607274296844e-06,
|
|
"loss": 0.0236,
|
|
"step": 435
|
|
},
|
|
{
|
|
"epoch": 1.4020908725371934,
|
|
"grad_norm": 0.634899377822876,
|
|
"learning_rate": 2.4514014710596467e-06,
|
|
"loss": 0.0205,
|
|
"step": 436
|
|
},
|
|
{
|
|
"epoch": 1.4053075995174908,
|
|
"grad_norm": 1.0875910520553589,
|
|
"learning_rate": 2.4272229998568576e-06,
|
|
"loss": 0.0367,
|
|
"step": 437
|
|
},
|
|
{
|
|
"epoch": 1.4085243264977885,
|
|
"grad_norm": 1.6455551385879517,
|
|
"learning_rate": 2.4031260802275623e-06,
|
|
"loss": 0.0271,
|
|
"step": 438
|
|
},
|
|
{
|
|
"epoch": 1.411741053478086,
|
|
"grad_norm": 0.7204722166061401,
|
|
"learning_rate": 2.3791114759930013e-06,
|
|
"loss": 0.0188,
|
|
"step": 439
|
|
},
|
|
{
|
|
"epoch": 1.4149577804583835,
|
|
"grad_norm": 1.0633268356323242,
|
|
"learning_rate": 2.3551799483651894e-06,
|
|
"loss": 0.0264,
|
|
"step": 440
|
|
},
|
|
{
|
|
"epoch": 1.4181745074386811,
|
|
"grad_norm": 0.6467992663383484,
|
|
"learning_rate": 2.331332255922784e-06,
|
|
"loss": 0.0184,
|
|
"step": 441
|
|
},
|
|
{
|
|
"epoch": 1.4213912344189787,
|
|
"grad_norm": 0.5554575324058533,
|
|
"learning_rate": 2.307569154587056e-06,
|
|
"loss": 0.0184,
|
|
"step": 442
|
|
},
|
|
{
|
|
"epoch": 1.4246079613992761,
|
|
"grad_norm": 0.6711392998695374,
|
|
"learning_rate": 2.283891397597908e-06,
|
|
"loss": 0.0206,
|
|
"step": 443
|
|
},
|
|
{
|
|
"epoch": 1.4278246883795738,
|
|
"grad_norm": 0.7576943635940552,
|
|
"learning_rate": 2.2602997354900075e-06,
|
|
"loss": 0.031,
|
|
"step": 444
|
|
},
|
|
{
|
|
"epoch": 1.4310414153598714,
|
|
"grad_norm": 0.6453703045845032,
|
|
"learning_rate": 2.236794916069007e-06,
|
|
"loss": 0.0171,
|
|
"step": 445
|
|
},
|
|
{
|
|
"epoch": 1.4342581423401688,
|
|
"grad_norm": 0.5706139802932739,
|
|
"learning_rate": 2.2133776843878185e-06,
|
|
"loss": 0.0191,
|
|
"step": 446
|
|
},
|
|
{
|
|
"epoch": 1.4374748693204664,
|
|
"grad_norm": 0.7611203789710999,
|
|
"learning_rate": 2.190048782723015e-06,
|
|
"loss": 0.0225,
|
|
"step": 447
|
|
},
|
|
{
|
|
"epoch": 1.440691596300764,
|
|
"grad_norm": 0.6822800040245056,
|
|
"learning_rate": 2.166808950551296e-06,
|
|
"loss": 0.0256,
|
|
"step": 448
|
|
},
|
|
{
|
|
"epoch": 1.4439083232810614,
|
|
"grad_norm": 0.798964262008667,
|
|
"learning_rate": 2.1436589245260375e-06,
|
|
"loss": 0.0236,
|
|
"step": 449
|
|
},
|
|
{
|
|
"epoch": 1.447125050261359,
|
|
"grad_norm": 0.8539571166038513,
|
|
"learning_rate": 2.120599438453968e-06,
|
|
"loss": 0.0241,
|
|
"step": 450
|
|
},
|
|
{
|
|
"epoch": 1.4503417772416567,
|
|
"grad_norm": 0.8464952707290649,
|
|
"learning_rate": 2.0976312232718763e-06,
|
|
"loss": 0.0282,
|
|
"step": 451
|
|
},
|
|
{
|
|
"epoch": 1.453558504221954,
|
|
"grad_norm": 0.7654972076416016,
|
|
"learning_rate": 2.074755007023461e-06,
|
|
"loss": 0.0278,
|
|
"step": 452
|
|
},
|
|
{
|
|
"epoch": 1.4567752312022517,
|
|
"grad_norm": 0.9112337231636047,
|
|
"learning_rate": 2.0519715148362585e-06,
|
|
"loss": 0.0176,
|
|
"step": 453
|
|
},
|
|
{
|
|
"epoch": 1.4599919581825493,
|
|
"grad_norm": 0.6856303215026855,
|
|
"learning_rate": 2.0292814688986375e-06,
|
|
"loss": 0.0224,
|
|
"step": 454
|
|
},
|
|
{
|
|
"epoch": 1.4632086851628467,
|
|
"grad_norm": 0.9206861257553101,
|
|
"learning_rate": 2.0066855884369246e-06,
|
|
"loss": 0.0312,
|
|
"step": 455
|
|
},
|
|
{
|
|
"epoch": 1.4664254121431444,
|
|
"grad_norm": 0.9237130880355835,
|
|
"learning_rate": 1.9841845896926022e-06,
|
|
"loss": 0.0256,
|
|
"step": 456
|
|
},
|
|
{
|
|
"epoch": 1.469642139123442,
|
|
"grad_norm": 0.5882164835929871,
|
|
"learning_rate": 1.961779185899597e-06,
|
|
"loss": 0.0224,
|
|
"step": 457
|
|
},
|
|
{
|
|
"epoch": 1.4728588661037394,
|
|
"grad_norm": 0.5949448347091675,
|
|
"learning_rate": 1.9394700872616856e-06,
|
|
"loss": 0.0197,
|
|
"step": 458
|
|
},
|
|
{
|
|
"epoch": 1.476075593084037,
|
|
"grad_norm": 0.5350119471549988,
|
|
"learning_rate": 1.9172580009299735e-06,
|
|
"loss": 0.0164,
|
|
"step": 459
|
|
},
|
|
{
|
|
"epoch": 1.4792923200643346,
|
|
"grad_norm": 1.140408992767334,
|
|
"learning_rate": 1.8951436309804766e-06,
|
|
"loss": 0.0344,
|
|
"step": 460
|
|
},
|
|
{
|
|
"epoch": 1.482509047044632,
|
|
"grad_norm": 0.7622132301330566,
|
|
"learning_rate": 1.8731276783918162e-06,
|
|
"loss": 0.0195,
|
|
"step": 461
|
|
},
|
|
{
|
|
"epoch": 1.4857257740249297,
|
|
"grad_norm": 1.5069661140441895,
|
|
"learning_rate": 1.8512108410229878e-06,
|
|
"loss": 0.0306,
|
|
"step": 462
|
|
},
|
|
{
|
|
"epoch": 1.4889425010052273,
|
|
"grad_norm": 0.7332004904747009,
|
|
"learning_rate": 1.8293938135912475e-06,
|
|
"loss": 0.0217,
|
|
"step": 463
|
|
},
|
|
{
|
|
"epoch": 1.4921592279855247,
|
|
"grad_norm": 0.7239487767219543,
|
|
"learning_rate": 1.8076772876500831e-06,
|
|
"loss": 0.0298,
|
|
"step": 464
|
|
},
|
|
{
|
|
"epoch": 1.4953759549658223,
|
|
"grad_norm": 0.8309620022773743,
|
|
"learning_rate": 1.7860619515673034e-06,
|
|
"loss": 0.0316,
|
|
"step": 465
|
|
},
|
|
{
|
|
"epoch": 1.49859268194612,
|
|
"grad_norm": 0.6418971419334412,
|
|
"learning_rate": 1.7645484905032129e-06,
|
|
"loss": 0.0182,
|
|
"step": 466
|
|
},
|
|
{
|
|
"epoch": 1.5018094089264173,
|
|
"grad_norm": 0.5187109112739563,
|
|
"learning_rate": 1.74313758638889e-06,
|
|
"loss": 0.025,
|
|
"step": 467
|
|
},
|
|
{
|
|
"epoch": 1.505026135906715,
|
|
"grad_norm": 0.9927024245262146,
|
|
"learning_rate": 1.7218299179045789e-06,
|
|
"loss": 0.0256,
|
|
"step": 468
|
|
},
|
|
{
|
|
"epoch": 1.5082428628870126,
|
|
"grad_norm": 0.6816557049751282,
|
|
"learning_rate": 1.7006261604581725e-06,
|
|
"loss": 0.0144,
|
|
"step": 469
|
|
},
|
|
{
|
|
"epoch": 1.51145958986731,
|
|
"grad_norm": 0.6767432689666748,
|
|
"learning_rate": 1.6795269861638041e-06,
|
|
"loss": 0.0256,
|
|
"step": 470
|
|
},
|
|
{
|
|
"epoch": 1.5146763168476074,
|
|
"grad_norm": 0.8080036044120789,
|
|
"learning_rate": 1.6585330638205454e-06,
|
|
"loss": 0.0283,
|
|
"step": 471
|
|
},
|
|
{
|
|
"epoch": 1.5178930438279052,
|
|
"grad_norm": 0.7607430219650269,
|
|
"learning_rate": 1.6376450588911985e-06,
|
|
"loss": 0.0223,
|
|
"step": 472
|
|
},
|
|
{
|
|
"epoch": 1.5211097708082026,
|
|
"grad_norm": 0.4729120433330536,
|
|
"learning_rate": 1.6168636334812126e-06,
|
|
"loss": 0.0142,
|
|
"step": 473
|
|
},
|
|
{
|
|
"epoch": 1.5243264977885,
|
|
"grad_norm": 0.5791823267936707,
|
|
"learning_rate": 1.5961894463176942e-06,
|
|
"loss": 0.0258,
|
|
"step": 474
|
|
},
|
|
{
|
|
"epoch": 1.527543224768798,
|
|
"grad_norm": 1.1532747745513916,
|
|
"learning_rate": 1.5756231527285181e-06,
|
|
"loss": 0.027,
|
|
"step": 475
|
|
},
|
|
{
|
|
"epoch": 1.5307599517490953,
|
|
"grad_norm": 0.7709406018257141,
|
|
"learning_rate": 1.555165404621567e-06,
|
|
"loss": 0.0273,
|
|
"step": 476
|
|
},
|
|
{
|
|
"epoch": 1.5339766787293927,
|
|
"grad_norm": 1.0912410020828247,
|
|
"learning_rate": 1.5348168504640631e-06,
|
|
"loss": 0.0282,
|
|
"step": 477
|
|
},
|
|
{
|
|
"epoch": 1.5371934057096905,
|
|
"grad_norm": 0.7054423689842224,
|
|
"learning_rate": 1.5145781352620054e-06,
|
|
"loss": 0.0273,
|
|
"step": 478
|
|
},
|
|
{
|
|
"epoch": 1.540410132689988,
|
|
"grad_norm": 0.46668004989624023,
|
|
"learning_rate": 1.4944499005397372e-06,
|
|
"loss": 0.0164,
|
|
"step": 479
|
|
},
|
|
{
|
|
"epoch": 1.5436268596702853,
|
|
"grad_norm": 0.7803506851196289,
|
|
"learning_rate": 1.4744327843196043e-06,
|
|
"loss": 0.0283,
|
|
"step": 480
|
|
},
|
|
{
|
|
"epoch": 1.5468435866505832,
|
|
"grad_norm": 0.7836430668830872,
|
|
"learning_rate": 1.4545274211017264e-06,
|
|
"loss": 0.0278,
|
|
"step": 481
|
|
},
|
|
{
|
|
"epoch": 1.5500603136308806,
|
|
"grad_norm": 1.152942419052124,
|
|
"learning_rate": 1.434734441843899e-06,
|
|
"loss": 0.0273,
|
|
"step": 482
|
|
},
|
|
{
|
|
"epoch": 1.553277040611178,
|
|
"grad_norm": 0.7189183831214905,
|
|
"learning_rate": 1.4150544739415755e-06,
|
|
"loss": 0.0238,
|
|
"step": 483
|
|
},
|
|
{
|
|
"epoch": 1.5564937675914756,
|
|
"grad_norm": 1.2334411144256592,
|
|
"learning_rate": 1.3954881412079945e-06,
|
|
"loss": 0.0296,
|
|
"step": 484
|
|
},
|
|
{
|
|
"epoch": 1.5597104945717732,
|
|
"grad_norm": 1.2712557315826416,
|
|
"learning_rate": 1.3760360638544012e-06,
|
|
"loss": 0.0258,
|
|
"step": 485
|
|
},
|
|
{
|
|
"epoch": 1.5629272215520706,
|
|
"grad_norm": 0.6514741778373718,
|
|
"learning_rate": 1.3566988584703817e-06,
|
|
"loss": 0.0256,
|
|
"step": 486
|
|
},
|
|
{
|
|
"epoch": 1.5661439485323683,
|
|
"grad_norm": 0.7159017324447632,
|
|
"learning_rate": 1.3374771380043306e-06,
|
|
"loss": 0.0221,
|
|
"step": 487
|
|
},
|
|
{
|
|
"epoch": 1.569360675512666,
|
|
"grad_norm": 0.7216795086860657,
|
|
"learning_rate": 1.3183715117440143e-06,
|
|
"loss": 0.0293,
|
|
"step": 488
|
|
},
|
|
{
|
|
"epoch": 1.5725774024929633,
|
|
"grad_norm": 0.5431943535804749,
|
|
"learning_rate": 1.2993825852972559e-06,
|
|
"loss": 0.0152,
|
|
"step": 489
|
|
},
|
|
{
|
|
"epoch": 1.575794129473261,
|
|
"grad_norm": 1.1157512664794922,
|
|
"learning_rate": 1.280510960572745e-06,
|
|
"loss": 0.0297,
|
|
"step": 490
|
|
},
|
|
{
|
|
"epoch": 1.5790108564535585,
|
|
"grad_norm": 0.8432955741882324,
|
|
"learning_rate": 1.2617572357609565e-06,
|
|
"loss": 0.023,
|
|
"step": 491
|
|
},
|
|
{
|
|
"epoch": 1.582227583433856,
|
|
"grad_norm": 0.8817299008369446,
|
|
"learning_rate": 1.2431220053151832e-06,
|
|
"loss": 0.0276,
|
|
"step": 492
|
|
},
|
|
{
|
|
"epoch": 1.5854443104141536,
|
|
"grad_norm": 0.6438712477684021,
|
|
"learning_rate": 1.2246058599327021e-06,
|
|
"loss": 0.0197,
|
|
"step": 493
|
|
},
|
|
{
|
|
"epoch": 1.5886610373944512,
|
|
"grad_norm": 0.947399377822876,
|
|
"learning_rate": 1.2062093865360458e-06,
|
|
"loss": 0.0266,
|
|
"step": 494
|
|
},
|
|
{
|
|
"epoch": 1.5918777643747486,
|
|
"grad_norm": 0.8782487511634827,
|
|
"learning_rate": 1.1879331682543972e-06,
|
|
"loss": 0.0146,
|
|
"step": 495
|
|
},
|
|
{
|
|
"epoch": 1.5950944913550462,
|
|
"grad_norm": 0.676629364490509,
|
|
"learning_rate": 1.1697777844051105e-06,
|
|
"loss": 0.0266,
|
|
"step": 496
|
|
},
|
|
{
|
|
"epoch": 1.5983112183353438,
|
|
"grad_norm": 1.0170975923538208,
|
|
"learning_rate": 1.1517438104753386e-06,
|
|
"loss": 0.0278,
|
|
"step": 497
|
|
},
|
|
{
|
|
"epoch": 1.6015279453156412,
|
|
"grad_norm": 0.8391772508621216,
|
|
"learning_rate": 1.1338318181038037e-06,
|
|
"loss": 0.0264,
|
|
"step": 498
|
|
},
|
|
{
|
|
"epoch": 1.6047446722959389,
|
|
"grad_norm": 0.6024744510650635,
|
|
"learning_rate": 1.1160423750626693e-06,
|
|
"loss": 0.0297,
|
|
"step": 499
|
|
},
|
|
{
|
|
"epoch": 1.6079613992762365,
|
|
"grad_norm": 0.5469645261764526,
|
|
"learning_rate": 1.0983760452395415e-06,
|
|
"loss": 0.0181,
|
|
"step": 500
|
|
},
|
|
{
|
|
"epoch": 1.611178126256534,
|
|
"grad_norm": 0.6368651986122131,
|
|
"learning_rate": 1.0808333886196038e-06,
|
|
"loss": 0.0203,
|
|
"step": 501
|
|
},
|
|
{
|
|
"epoch": 1.6143948532368315,
|
|
"grad_norm": 0.6552625894546509,
|
|
"learning_rate": 1.063414961267859e-06,
|
|
"loss": 0.0243,
|
|
"step": 502
|
|
},
|
|
{
|
|
"epoch": 1.6176115802171291,
|
|
"grad_norm": 0.5399875640869141,
|
|
"learning_rate": 1.046121315311508e-06,
|
|
"loss": 0.0201,
|
|
"step": 503
|
|
},
|
|
{
|
|
"epoch": 1.6208283071974265,
|
|
"grad_norm": 0.7226431369781494,
|
|
"learning_rate": 1.02895299892244e-06,
|
|
"loss": 0.0269,
|
|
"step": 504
|
|
},
|
|
{
|
|
"epoch": 1.6240450341777242,
|
|
"grad_norm": 0.7085054516792297,
|
|
"learning_rate": 1.01191055629987e-06,
|
|
"loss": 0.0253,
|
|
"step": 505
|
|
},
|
|
{
|
|
"epoch": 1.6272617611580218,
|
|
"grad_norm": 0.7027333974838257,
|
|
"learning_rate": 9.949945276530782e-07,
|
|
"loss": 0.0202,
|
|
"step": 506
|
|
},
|
|
{
|
|
"epoch": 1.6304784881383192,
|
|
"grad_norm": 0.669688880443573,
|
|
"learning_rate": 9.782054491842879e-07,
|
|
"loss": 0.0184,
|
|
"step": 507
|
|
},
|
|
{
|
|
"epoch": 1.6336952151186168,
|
|
"grad_norm": 0.7410572171211243,
|
|
"learning_rate": 9.615438530716753e-07,
|
|
"loss": 0.0144,
|
|
"step": 508
|
|
},
|
|
{
|
|
"epoch": 1.6369119420989144,
|
|
"grad_norm": 0.8965449333190918,
|
|
"learning_rate": 9.450102674524952e-07,
|
|
"loss": 0.0206,
|
|
"step": 509
|
|
},
|
|
{
|
|
"epoch": 1.6401286690792118,
|
|
"grad_norm": 0.8675963878631592,
|
|
"learning_rate": 9.286052164063369e-07,
|
|
"loss": 0.0267,
|
|
"step": 510
|
|
},
|
|
{
|
|
"epoch": 1.6433453960595095,
|
|
"grad_norm": 0.9799548387527466,
|
|
"learning_rate": 9.123292199385247e-07,
|
|
"loss": 0.0334,
|
|
"step": 511
|
|
},
|
|
{
|
|
"epoch": 1.646562123039807,
|
|
"grad_norm": 0.5996445417404175,
|
|
"learning_rate": 8.961827939636198e-07,
|
|
"loss": 0.023,
|
|
"step": 512
|
|
},
|
|
{
|
|
"epoch": 1.6497788500201045,
|
|
"grad_norm": 0.6122687458992004,
|
|
"learning_rate": 8.801664502890722e-07,
|
|
"loss": 0.0214,
|
|
"step": 513
|
|
},
|
|
{
|
|
"epoch": 1.6529955770004021,
|
|
"grad_norm": 0.6234684586524963,
|
|
"learning_rate": 8.64280696599008e-07,
|
|
"loss": 0.0215,
|
|
"step": 514
|
|
},
|
|
{
|
|
"epoch": 1.6562123039806997,
|
|
"grad_norm": 0.5246209502220154,
|
|
"learning_rate": 8.485260364381187e-07,
|
|
"loss": 0.0186,
|
|
"step": 515
|
|
},
|
|
{
|
|
"epoch": 1.6594290309609971,
|
|
"grad_norm": 1.306154489517212,
|
|
"learning_rate": 8.329029691957124e-07,
|
|
"loss": 0.0257,
|
|
"step": 516
|
|
},
|
|
{
|
|
"epoch": 1.6626457579412948,
|
|
"grad_norm": 0.8695035576820374,
|
|
"learning_rate": 8.17411990089883e-07,
|
|
"loss": 0.0389,
|
|
"step": 517
|
|
},
|
|
{
|
|
"epoch": 1.6658624849215924,
|
|
"grad_norm": 0.5970683693885803,
|
|
"learning_rate": 8.02053590151805e-07,
|
|
"loss": 0.0281,
|
|
"step": 518
|
|
},
|
|
{
|
|
"epoch": 1.6690792119018898,
|
|
"grad_norm": 0.7056528329849243,
|
|
"learning_rate": 7.868282562101819e-07,
|
|
"loss": 0.0212,
|
|
"step": 519
|
|
},
|
|
{
|
|
"epoch": 1.6722959388821874,
|
|
"grad_norm": 0.8744556307792664,
|
|
"learning_rate": 7.717364708758024e-07,
|
|
"loss": 0.0265,
|
|
"step": 520
|
|
},
|
|
{
|
|
"epoch": 1.675512665862485,
|
|
"grad_norm": 0.8215451240539551,
|
|
"learning_rate": 7.567787125262449e-07,
|
|
"loss": 0.0197,
|
|
"step": 521
|
|
},
|
|
{
|
|
"epoch": 1.6787293928427824,
|
|
"grad_norm": 1.1487066745758057,
|
|
"learning_rate": 7.41955455290726e-07,
|
|
"loss": 0.03,
|
|
"step": 522
|
|
},
|
|
{
|
|
"epoch": 1.68194611982308,
|
|
"grad_norm": 0.7303391098976135,
|
|
"learning_rate": 7.27267169035053e-07,
|
|
"loss": 0.0253,
|
|
"step": 523
|
|
},
|
|
{
|
|
"epoch": 1.6851628468033777,
|
|
"grad_norm": 0.5748582482337952,
|
|
"learning_rate": 7.127143193467445e-07,
|
|
"loss": 0.0209,
|
|
"step": 524
|
|
},
|
|
{
|
|
"epoch": 1.688379573783675,
|
|
"grad_norm": 1.0102810859680176,
|
|
"learning_rate": 6.982973675202676e-07,
|
|
"loss": 0.0263,
|
|
"step": 525
|
|
},
|
|
{
|
|
"epoch": 1.6915963007639725,
|
|
"grad_norm": 0.7919561862945557,
|
|
"learning_rate": 6.840167705424106e-07,
|
|
"loss": 0.0268,
|
|
"step": 526
|
|
},
|
|
{
|
|
"epoch": 1.6948130277442703,
|
|
"grad_norm": 0.792404294013977,
|
|
"learning_rate": 6.698729810778065e-07,
|
|
"loss": 0.0158,
|
|
"step": 527
|
|
},
|
|
{
|
|
"epoch": 1.6980297547245677,
|
|
"grad_norm": 1.081179141998291,
|
|
"learning_rate": 6.558664474545817e-07,
|
|
"loss": 0.0312,
|
|
"step": 528
|
|
},
|
|
{
|
|
"epoch": 1.7012464817048651,
|
|
"grad_norm": 0.6410866975784302,
|
|
"learning_rate": 6.419976136501377e-07,
|
|
"loss": 0.0188,
|
|
"step": 529
|
|
},
|
|
{
|
|
"epoch": 1.704463208685163,
|
|
"grad_norm": 0.5669673085212708,
|
|
"learning_rate": 6.282669192770896e-07,
|
|
"loss": 0.0164,
|
|
"step": 530
|
|
},
|
|
{
|
|
"epoch": 1.7076799356654604,
|
|
"grad_norm": 0.6292052268981934,
|
|
"learning_rate": 6.146747995693225e-07,
|
|
"loss": 0.0251,
|
|
"step": 531
|
|
},
|
|
{
|
|
"epoch": 1.7108966626457578,
|
|
"grad_norm": 0.8831068277359009,
|
|
"learning_rate": 6.012216853682001e-07,
|
|
"loss": 0.0271,
|
|
"step": 532
|
|
},
|
|
{
|
|
"epoch": 1.7141133896260556,
|
|
"grad_norm": 0.8742364048957825,
|
|
"learning_rate": 5.879080031089047e-07,
|
|
"loss": 0.0297,
|
|
"step": 533
|
|
},
|
|
{
|
|
"epoch": 1.717330116606353,
|
|
"grad_norm": 0.6837791800498962,
|
|
"learning_rate": 5.747341748069229e-07,
|
|
"loss": 0.0184,
|
|
"step": 534
|
|
},
|
|
{
|
|
"epoch": 1.7205468435866504,
|
|
"grad_norm": 0.7482956647872925,
|
|
"learning_rate": 5.617006180446688e-07,
|
|
"loss": 0.0237,
|
|
"step": 535
|
|
},
|
|
{
|
|
"epoch": 1.7237635705669483,
|
|
"grad_norm": 0.5687646269798279,
|
|
"learning_rate": 5.488077459582425e-07,
|
|
"loss": 0.0219,
|
|
"step": 536
|
|
},
|
|
{
|
|
"epoch": 1.7269802975472457,
|
|
"grad_norm": 0.5424864292144775,
|
|
"learning_rate": 5.360559672243421e-07,
|
|
"loss": 0.0228,
|
|
"step": 537
|
|
},
|
|
{
|
|
"epoch": 1.730197024527543,
|
|
"grad_norm": 0.6900858879089355,
|
|
"learning_rate": 5.234456860473042e-07,
|
|
"loss": 0.0227,
|
|
"step": 538
|
|
},
|
|
{
|
|
"epoch": 1.7334137515078407,
|
|
"grad_norm": 0.5527268648147583,
|
|
"learning_rate": 5.109773021462921e-07,
|
|
"loss": 0.0156,
|
|
"step": 539
|
|
},
|
|
{
|
|
"epoch": 1.7366304784881383,
|
|
"grad_norm": 0.7617378234863281,
|
|
"learning_rate": 4.986512107426283e-07,
|
|
"loss": 0.0308,
|
|
"step": 540
|
|
},
|
|
{
|
|
"epoch": 1.7398472054684357,
|
|
"grad_norm": 0.8059920072555542,
|
|
"learning_rate": 4.864678025472635e-07,
|
|
"loss": 0.0119,
|
|
"step": 541
|
|
},
|
|
{
|
|
"epoch": 1.7430639324487334,
|
|
"grad_norm": 0.8831843137741089,
|
|
"learning_rate": 4.7442746374839363e-07,
|
|
"loss": 0.0267,
|
|
"step": 542
|
|
},
|
|
{
|
|
"epoch": 1.746280659429031,
|
|
"grad_norm": 0.5986456274986267,
|
|
"learning_rate": 4.625305759992205e-07,
|
|
"loss": 0.0185,
|
|
"step": 543
|
|
},
|
|
{
|
|
"epoch": 1.7494973864093284,
|
|
"grad_norm": 0.9453976154327393,
|
|
"learning_rate": 4.50777516405847e-07,
|
|
"loss": 0.0244,
|
|
"step": 544
|
|
},
|
|
{
|
|
"epoch": 1.752714113389626,
|
|
"grad_norm": 0.7596896290779114,
|
|
"learning_rate": 4.3916865751533313e-07,
|
|
"loss": 0.024,
|
|
"step": 545
|
|
},
|
|
{
|
|
"epoch": 1.7559308403699236,
|
|
"grad_norm": 0.9080878496170044,
|
|
"learning_rate": 4.2770436730388166e-07,
|
|
"loss": 0.0281,
|
|
"step": 546
|
|
},
|
|
{
|
|
"epoch": 1.759147567350221,
|
|
"grad_norm": 1.2852977514266968,
|
|
"learning_rate": 4.163850091651717e-07,
|
|
"loss": 0.0345,
|
|
"step": 547
|
|
},
|
|
{
|
|
"epoch": 1.7623642943305187,
|
|
"grad_norm": 0.9868035912513733,
|
|
"learning_rate": 4.05210941898847e-07,
|
|
"loss": 0.0203,
|
|
"step": 548
|
|
},
|
|
{
|
|
"epoch": 1.7655810213108163,
|
|
"grad_norm": 0.749136745929718,
|
|
"learning_rate": 3.941825196991378e-07,
|
|
"loss": 0.0249,
|
|
"step": 549
|
|
},
|
|
{
|
|
"epoch": 1.7687977482911137,
|
|
"grad_norm": 0.9028546810150146,
|
|
"learning_rate": 3.8330009214363197e-07,
|
|
"loss": 0.0218,
|
|
"step": 550
|
|
},
|
|
{
|
|
"epoch": 1.7720144752714113,
|
|
"grad_norm": 0.8566029667854309,
|
|
"learning_rate": 3.725640041822026e-07,
|
|
"loss": 0.0251,
|
|
"step": 551
|
|
},
|
|
{
|
|
"epoch": 1.775231202251709,
|
|
"grad_norm": 0.6764525771141052,
|
|
"learning_rate": 3.619745961260623e-07,
|
|
"loss": 0.0255,
|
|
"step": 552
|
|
},
|
|
{
|
|
"epoch": 1.7784479292320063,
|
|
"grad_norm": 0.8676577210426331,
|
|
"learning_rate": 3.5153220363698225e-07,
|
|
"loss": 0.0233,
|
|
"step": 553
|
|
},
|
|
{
|
|
"epoch": 1.781664656212304,
|
|
"grad_norm": 0.7067648768424988,
|
|
"learning_rate": 3.4123715771665786e-07,
|
|
"loss": 0.0211,
|
|
"step": 554
|
|
},
|
|
{
|
|
"epoch": 1.7848813831926016,
|
|
"grad_norm": 0.6299604773521423,
|
|
"learning_rate": 3.310897846962041e-07,
|
|
"loss": 0.0158,
|
|
"step": 555
|
|
},
|
|
{
|
|
"epoch": 1.788098110172899,
|
|
"grad_norm": 0.5761677026748657,
|
|
"learning_rate": 3.2109040622582186e-07,
|
|
"loss": 0.0129,
|
|
"step": 556
|
|
},
|
|
{
|
|
"epoch": 1.7913148371531966,
|
|
"grad_norm": 0.7623018622398376,
|
|
"learning_rate": 3.112393392645985e-07,
|
|
"loss": 0.0241,
|
|
"step": 557
|
|
},
|
|
{
|
|
"epoch": 1.7945315641334942,
|
|
"grad_norm": 0.8402174711227417,
|
|
"learning_rate": 3.015368960704584e-07,
|
|
"loss": 0.026,
|
|
"step": 558
|
|
},
|
|
{
|
|
"epoch": 1.7977482911137916,
|
|
"grad_norm": 0.7599536776542664,
|
|
"learning_rate": 2.919833841902714e-07,
|
|
"loss": 0.0229,
|
|
"step": 559
|
|
},
|
|
{
|
|
"epoch": 1.8009650180940893,
|
|
"grad_norm": 0.5555005073547363,
|
|
"learning_rate": 2.8257910645009935e-07,
|
|
"loss": 0.025,
|
|
"step": 560
|
|
},
|
|
{
|
|
"epoch": 1.8041817450743869,
|
|
"grad_norm": 0.9722334742546082,
|
|
"learning_rate": 2.733243609455971e-07,
|
|
"loss": 0.0248,
|
|
"step": 561
|
|
},
|
|
{
|
|
"epoch": 1.8073984720546843,
|
|
"grad_norm": 0.47031423449516296,
|
|
"learning_rate": 2.6421944103256657e-07,
|
|
"loss": 0.0118,
|
|
"step": 562
|
|
},
|
|
{
|
|
"epoch": 1.810615199034982,
|
|
"grad_norm": 0.6891758441925049,
|
|
"learning_rate": 2.5526463531765467e-07,
|
|
"loss": 0.0272,
|
|
"step": 563
|
|
},
|
|
{
|
|
"epoch": 1.8138319260152795,
|
|
"grad_norm": 0.5683274269104004,
|
|
"learning_rate": 2.4646022764920843e-07,
|
|
"loss": 0.0181,
|
|
"step": 564
|
|
},
|
|
{
|
|
"epoch": 1.817048652995577,
|
|
"grad_norm": 0.5917408466339111,
|
|
"learning_rate": 2.3780649710827552e-07,
|
|
"loss": 0.0223,
|
|
"step": 565
|
|
},
|
|
{
|
|
"epoch": 1.8202653799758746,
|
|
"grad_norm": 0.6025788187980652,
|
|
"learning_rate": 2.2930371799975593e-07,
|
|
"loss": 0.0209,
|
|
"step": 566
|
|
},
|
|
{
|
|
"epoch": 1.8234821069561722,
|
|
"grad_norm": 0.6595741510391235,
|
|
"learning_rate": 2.20952159843712e-07,
|
|
"loss": 0.0197,
|
|
"step": 567
|
|
},
|
|
{
|
|
"epoch": 1.8266988339364696,
|
|
"grad_norm": 0.5931892395019531,
|
|
"learning_rate": 2.1275208736682262e-07,
|
|
"loss": 0.0149,
|
|
"step": 568
|
|
},
|
|
{
|
|
"epoch": 1.8299155609167672,
|
|
"grad_norm": 0.45539724826812744,
|
|
"learning_rate": 2.0470376049398944e-07,
|
|
"loss": 0.0132,
|
|
"step": 569
|
|
},
|
|
{
|
|
"epoch": 1.8331322878970648,
|
|
"grad_norm": 0.7427828311920166,
|
|
"learning_rate": 1.9680743434010385e-07,
|
|
"loss": 0.0201,
|
|
"step": 570
|
|
},
|
|
{
|
|
"epoch": 1.8363490148773622,
|
|
"grad_norm": 0.7957307696342468,
|
|
"learning_rate": 1.8906335920195418e-07,
|
|
"loss": 0.0172,
|
|
"step": 571
|
|
},
|
|
{
|
|
"epoch": 1.8395657418576599,
|
|
"grad_norm": 0.6128239035606384,
|
|
"learning_rate": 1.814717805502958e-07,
|
|
"loss": 0.023,
|
|
"step": 572
|
|
},
|
|
{
|
|
"epoch": 1.8427824688379575,
|
|
"grad_norm": 0.6324180364608765,
|
|
"learning_rate": 1.7403293902206851e-07,
|
|
"loss": 0.0231,
|
|
"step": 573
|
|
},
|
|
{
|
|
"epoch": 1.845999195818255,
|
|
"grad_norm": 0.781023383140564,
|
|
"learning_rate": 1.667470704127694e-07,
|
|
"loss": 0.0323,
|
|
"step": 574
|
|
},
|
|
{
|
|
"epoch": 1.8492159227985525,
|
|
"grad_norm": 0.5555590391159058,
|
|
"learning_rate": 1.5961440566897913e-07,
|
|
"loss": 0.0237,
|
|
"step": 575
|
|
},
|
|
{
|
|
"epoch": 1.8524326497788501,
|
|
"grad_norm": 0.8592231869697571,
|
|
"learning_rate": 1.5263517088103862e-07,
|
|
"loss": 0.0232,
|
|
"step": 576
|
|
},
|
|
{
|
|
"epoch": 1.8556493767591475,
|
|
"grad_norm": 0.6279895901679993,
|
|
"learning_rate": 1.4580958727588746e-07,
|
|
"loss": 0.0154,
|
|
"step": 577
|
|
},
|
|
{
|
|
"epoch": 1.8588661037394452,
|
|
"grad_norm": 0.7780999541282654,
|
|
"learning_rate": 1.3913787121004717e-07,
|
|
"loss": 0.0255,
|
|
"step": 578
|
|
},
|
|
{
|
|
"epoch": 1.8620828307197428,
|
|
"grad_norm": 0.9826112985610962,
|
|
"learning_rate": 1.3262023416276414e-07,
|
|
"loss": 0.0293,
|
|
"step": 579
|
|
},
|
|
{
|
|
"epoch": 1.8652995577000402,
|
|
"grad_norm": 0.7337117791175842,
|
|
"learning_rate": 1.2625688272930925e-07,
|
|
"loss": 0.0244,
|
|
"step": 580
|
|
},
|
|
{
|
|
"epoch": 1.8685162846803376,
|
|
"grad_norm": 0.7108997106552124,
|
|
"learning_rate": 1.2004801861442373e-07,
|
|
"loss": 0.0227,
|
|
"step": 581
|
|
},
|
|
{
|
|
"epoch": 1.8717330116606354,
|
|
"grad_norm": 0.6001144647598267,
|
|
"learning_rate": 1.1399383862592928e-07,
|
|
"loss": 0.0242,
|
|
"step": 582
|
|
},
|
|
{
|
|
"epoch": 1.8749497386409328,
|
|
"grad_norm": 1.1188668012619019,
|
|
"learning_rate": 1.0809453466849029e-07,
|
|
"loss": 0.0277,
|
|
"step": 583
|
|
},
|
|
{
|
|
"epoch": 1.8781664656212302,
|
|
"grad_norm": 0.7146681547164917,
|
|
"learning_rate": 1.0235029373752758e-07,
|
|
"loss": 0.0218,
|
|
"step": 584
|
|
},
|
|
{
|
|
"epoch": 1.881383192601528,
|
|
"grad_norm": 0.6120503544807434,
|
|
"learning_rate": 9.676129791329481e-08,
|
|
"loss": 0.0232,
|
|
"step": 585
|
|
},
|
|
{
|
|
"epoch": 1.8845999195818255,
|
|
"grad_norm": 0.69977205991745,
|
|
"learning_rate": 9.132772435510362e-08,
|
|
"loss": 0.0255,
|
|
"step": 586
|
|
},
|
|
{
|
|
"epoch": 1.887816646562123,
|
|
"grad_norm": 0.9345167279243469,
|
|
"learning_rate": 8.604974529571042e-08,
|
|
"loss": 0.0368,
|
|
"step": 587
|
|
},
|
|
{
|
|
"epoch": 1.8910333735424207,
|
|
"grad_norm": 0.7977281212806702,
|
|
"learning_rate": 8.092752803585513e-08,
|
|
"loss": 0.0234,
|
|
"step": 588
|
|
},
|
|
{
|
|
"epoch": 1.8942501005227181,
|
|
"grad_norm": 0.7188998460769653,
|
|
"learning_rate": 7.59612349389599e-08,
|
|
"loss": 0.0222,
|
|
"step": 589
|
|
},
|
|
{
|
|
"epoch": 1.8974668275030155,
|
|
"grad_norm": 1.3486685752868652,
|
|
"learning_rate": 7.115102342598101e-08,
|
|
"loss": 0.0347,
|
|
"step": 590
|
|
},
|
|
{
|
|
"epoch": 1.9006835544833134,
|
|
"grad_norm": 0.7500903010368347,
|
|
"learning_rate": 6.649704597042061e-08,
|
|
"loss": 0.022,
|
|
"step": 591
|
|
},
|
|
{
|
|
"epoch": 1.9039002814636108,
|
|
"grad_norm": 0.5467425584793091,
|
|
"learning_rate": 6.199945009349173e-08,
|
|
"loss": 0.0195,
|
|
"step": 592
|
|
},
|
|
{
|
|
"epoch": 1.9071170084439082,
|
|
"grad_norm": 0.7178412079811096,
|
|
"learning_rate": 5.7658378359443104e-08,
|
|
"loss": 0.0307,
|
|
"step": 593
|
|
},
|
|
{
|
|
"epoch": 1.910333735424206,
|
|
"grad_norm": 0.8701561093330383,
|
|
"learning_rate": 5.3473968371040575e-08,
|
|
"loss": 0.0206,
|
|
"step": 594
|
|
},
|
|
{
|
|
"epoch": 1.9135504624045034,
|
|
"grad_norm": 0.7027232050895691,
|
|
"learning_rate": 4.944635276520393e-08,
|
|
"loss": 0.0231,
|
|
"step": 595
|
|
},
|
|
{
|
|
"epoch": 1.9167671893848008,
|
|
"grad_norm": 0.711676299571991,
|
|
"learning_rate": 4.55756592088058e-08,
|
|
"loss": 0.0219,
|
|
"step": 596
|
|
},
|
|
{
|
|
"epoch": 1.9199839163650985,
|
|
"grad_norm": 0.8163571357727051,
|
|
"learning_rate": 4.186201039462046e-08,
|
|
"loss": 0.0243,
|
|
"step": 597
|
|
},
|
|
{
|
|
"epoch": 1.923200643345396,
|
|
"grad_norm": 0.781928539276123,
|
|
"learning_rate": 3.8305524037438035e-08,
|
|
"loss": 0.0269,
|
|
"step": 598
|
|
},
|
|
{
|
|
"epoch": 1.9264173703256935,
|
|
"grad_norm": 0.7710465788841248,
|
|
"learning_rate": 3.4906312870331973e-08,
|
|
"loss": 0.0172,
|
|
"step": 599
|
|
},
|
|
{
|
|
"epoch": 1.9296340973059911,
|
|
"grad_norm": 0.9015802145004272,
|
|
"learning_rate": 3.166448464108629e-08,
|
|
"loss": 0.0302,
|
|
"step": 600
|
|
},
|
|
{
|
|
"epoch": 1.9328508242862887,
|
|
"grad_norm": 0.857291042804718,
|
|
"learning_rate": 2.8580142108778354e-08,
|
|
"loss": 0.0182,
|
|
"step": 601
|
|
},
|
|
{
|
|
"epoch": 1.9360675512665861,
|
|
"grad_norm": 0.8512669801712036,
|
|
"learning_rate": 2.5653383040524228e-08,
|
|
"loss": 0.0299,
|
|
"step": 602
|
|
},
|
|
{
|
|
"epoch": 1.9392842782468838,
|
|
"grad_norm": 0.8255121111869812,
|
|
"learning_rate": 2.2884300208378395e-08,
|
|
"loss": 0.0278,
|
|
"step": 603
|
|
},
|
|
{
|
|
"epoch": 1.9425010052271814,
|
|
"grad_norm": 0.5912429690361023,
|
|
"learning_rate": 2.0272981386393332e-08,
|
|
"loss": 0.0207,
|
|
"step": 604
|
|
},
|
|
{
|
|
"epoch": 1.9457177322074788,
|
|
"grad_norm": 0.6523793935775757,
|
|
"learning_rate": 1.781950934783505e-08,
|
|
"loss": 0.0177,
|
|
"step": 605
|
|
},
|
|
{
|
|
"epoch": 1.9489344591877764,
|
|
"grad_norm": 0.7320718169212341,
|
|
"learning_rate": 1.552396186256411e-08,
|
|
"loss": 0.0255,
|
|
"step": 606
|
|
},
|
|
{
|
|
"epoch": 1.952151186168074,
|
|
"grad_norm": 0.6994187235832214,
|
|
"learning_rate": 1.3386411694565894e-08,
|
|
"loss": 0.0278,
|
|
"step": 607
|
|
},
|
|
{
|
|
"epoch": 1.9553679131483714,
|
|
"grad_norm": 0.5923306941986084,
|
|
"learning_rate": 1.1406926599646373e-08,
|
|
"loss": 0.0192,
|
|
"step": 608
|
|
},
|
|
{
|
|
"epoch": 1.958584640128669,
|
|
"grad_norm": 0.7130267024040222,
|
|
"learning_rate": 9.585569323284915e-09,
|
|
"loss": 0.0265,
|
|
"step": 609
|
|
},
|
|
{
|
|
"epoch": 1.9618013671089667,
|
|
"grad_norm": 0.9426466226577759,
|
|
"learning_rate": 7.922397598642551e-09,
|
|
"loss": 0.0357,
|
|
"step": 610
|
|
},
|
|
{
|
|
"epoch": 1.965018094089264,
|
|
"grad_norm": 0.6841176748275757,
|
|
"learning_rate": 6.417464144736208e-09,
|
|
"loss": 0.0232,
|
|
"step": 611
|
|
},
|
|
{
|
|
"epoch": 1.9682348210695617,
|
|
"grad_norm": 0.6261034607887268,
|
|
"learning_rate": 5.0708166647628345e-09,
|
|
"loss": 0.0205,
|
|
"step": 612
|
|
},
|
|
{
|
|
"epoch": 1.9714515480498593,
|
|
"grad_norm": 0.7374976873397827,
|
|
"learning_rate": 3.88249784459227e-09,
|
|
"loss": 0.0193,
|
|
"step": 613
|
|
},
|
|
{
|
|
"epoch": 1.9746682750301567,
|
|
"grad_norm": 0.676289975643158,
|
|
"learning_rate": 2.8525453514099966e-09,
|
|
"loss": 0.0177,
|
|
"step": 614
|
|
},
|
|
{
|
|
"epoch": 1.9778850020104544,
|
|
"grad_norm": 0.8624617457389832,
|
|
"learning_rate": 1.980991832524759e-09,
|
|
"loss": 0.0385,
|
|
"step": 615
|
|
},
|
|
{
|
|
"epoch": 1.981101728990752,
|
|
"grad_norm": 0.6072701811790466,
|
|
"learning_rate": 1.2678649143349485e-09,
|
|
"loss": 0.0219,
|
|
"step": 616
|
|
},
|
|
{
|
|
"epoch": 1.9843184559710494,
|
|
"grad_norm": 0.620798647403717,
|
|
"learning_rate": 7.131872014509711e-10,
|
|
"loss": 0.0211,
|
|
"step": 617
|
|
},
|
|
{
|
|
"epoch": 1.987535182951347,
|
|
"grad_norm": 0.5879101753234863,
|
|
"learning_rate": 3.1697627597970794e-10,
|
|
"loss": 0.0169,
|
|
"step": 618
|
|
},
|
|
{
|
|
"epoch": 1.9907519099316446,
|
|
"grad_norm": 1.4460906982421875,
|
|
"learning_rate": 7.924469696718451e-11,
|
|
"loss": 0.0259,
|
|
"step": 619
|
|
},
|
|
{
|
|
"epoch": 1.993968636911942,
|
|
"grad_norm": 0.9200376868247986,
|
|
"learning_rate": 0.0,
|
|
"loss": 0.0252,
|
|
"step": 620
|
|
},
|
|
{
|
|
"epoch": 1.993968636911942,
|
|
"step": 620,
|
|
"total_flos": 7.073469349725471e+17,
|
|
"train_loss": 0.04489339354387935,
|
|
"train_runtime": 2537.7411,
|
|
"train_samples_per_second": 125.415,
|
|
"train_steps_per_second": 0.244
|
|
}
|
|
],
|
|
"logging_steps": 1.0,
|
|
"max_steps": 620,
|
|
"num_input_tokens_seen": 0,
|
|
"num_train_epochs": 2,
|
|
"save_steps": 500,
|
|
"stateful_callbacks": {
|
|
"TrainerControl": {
|
|
"args": {
|
|
"should_epoch_stop": false,
|
|
"should_evaluate": false,
|
|
"should_log": false,
|
|
"should_save": true,
|
|
"should_training_stop": true
|
|
},
|
|
"attributes": {}
|
|
}
|
|
},
|
|
"total_flos": 7.073469349725471e+17,
|
|
"train_batch_size": 8,
|
|
"trial_name": null,
|
|
"trial_params": null
|
|
}
|