1673 lines
40 KiB
JSON
1673 lines
40 KiB
JSON
{
|
|
"best_global_step": null,
|
|
"best_metric": null,
|
|
"best_model_checkpoint": null,
|
|
"epoch": 1.0,
|
|
"eval_steps": 500,
|
|
"global_step": 234,
|
|
"is_hyper_param_search": false,
|
|
"is_local_process_zero": true,
|
|
"is_world_process_zero": true,
|
|
"log_history": [
|
|
{
|
|
"epoch": 0.004282082162451492,
|
|
"grad_norm": 2.3133654594421387,
|
|
"learning_rate": 0.0,
|
|
"loss": 0.9749,
|
|
"step": 1
|
|
},
|
|
{
|
|
"epoch": 0.008564164324902984,
|
|
"grad_norm": 1.8335061073303223,
|
|
"learning_rate": 3.3333333333333333e-06,
|
|
"loss": 0.7648,
|
|
"step": 2
|
|
},
|
|
{
|
|
"epoch": 0.012846246487354477,
|
|
"grad_norm": 2.2353498935699463,
|
|
"learning_rate": 6.666666666666667e-06,
|
|
"loss": 0.9076,
|
|
"step": 3
|
|
},
|
|
{
|
|
"epoch": 0.017128328649805968,
|
|
"grad_norm": 2.1200685501098633,
|
|
"learning_rate": 1e-05,
|
|
"loss": 0.8221,
|
|
"step": 4
|
|
},
|
|
{
|
|
"epoch": 0.02141041081225746,
|
|
"grad_norm": 2.185300350189209,
|
|
"learning_rate": 1.3333333333333333e-05,
|
|
"loss": 0.8972,
|
|
"step": 5
|
|
},
|
|
{
|
|
"epoch": 0.025692492974708953,
|
|
"grad_norm": 2.383875608444214,
|
|
"learning_rate": 1.6666666666666667e-05,
|
|
"loss": 0.8918,
|
|
"step": 6
|
|
},
|
|
{
|
|
"epoch": 0.029974575137160443,
|
|
"grad_norm": 2.272465467453003,
|
|
"learning_rate": 2e-05,
|
|
"loss": 0.803,
|
|
"step": 7
|
|
},
|
|
{
|
|
"epoch": 0.034256657299611935,
|
|
"grad_norm": 1.9353933334350586,
|
|
"learning_rate": 2.3333333333333336e-05,
|
|
"loss": 0.5844,
|
|
"step": 8
|
|
},
|
|
{
|
|
"epoch": 0.03853873946206343,
|
|
"grad_norm": 1.7865619659423828,
|
|
"learning_rate": 2.6666666666666667e-05,
|
|
"loss": 0.5373,
|
|
"step": 9
|
|
},
|
|
{
|
|
"epoch": 0.04282082162451492,
|
|
"grad_norm": 1.5210086107254028,
|
|
"learning_rate": 3e-05,
|
|
"loss": 0.4767,
|
|
"step": 10
|
|
},
|
|
{
|
|
"epoch": 0.04710290378696641,
|
|
"grad_norm": 1.1693248748779297,
|
|
"learning_rate": 3.3333333333333335e-05,
|
|
"loss": 0.3285,
|
|
"step": 11
|
|
},
|
|
{
|
|
"epoch": 0.051384985949417906,
|
|
"grad_norm": 0.8487532138824463,
|
|
"learning_rate": 3.6666666666666666e-05,
|
|
"loss": 0.2321,
|
|
"step": 12
|
|
},
|
|
{
|
|
"epoch": 0.055667068111869396,
|
|
"grad_norm": 0.8226314187049866,
|
|
"learning_rate": 4e-05,
|
|
"loss": 0.2117,
|
|
"step": 13
|
|
},
|
|
{
|
|
"epoch": 0.059949150274320885,
|
|
"grad_norm": 0.7278441190719604,
|
|
"learning_rate": 4.3333333333333334e-05,
|
|
"loss": 0.1951,
|
|
"step": 14
|
|
},
|
|
{
|
|
"epoch": 0.06423123243677238,
|
|
"grad_norm": 0.6208946704864502,
|
|
"learning_rate": 4.666666666666667e-05,
|
|
"loss": 0.1739,
|
|
"step": 15
|
|
},
|
|
{
|
|
"epoch": 0.06851331459922387,
|
|
"grad_norm": 0.44936591386795044,
|
|
"learning_rate": 5e-05,
|
|
"loss": 0.1144,
|
|
"step": 16
|
|
},
|
|
{
|
|
"epoch": 0.07279539676167536,
|
|
"grad_norm": 0.42960935831069946,
|
|
"learning_rate": 5.333333333333333e-05,
|
|
"loss": 0.1089,
|
|
"step": 17
|
|
},
|
|
{
|
|
"epoch": 0.07707747892412686,
|
|
"grad_norm": 0.4222032427787781,
|
|
"learning_rate": 5.666666666666667e-05,
|
|
"loss": 0.1103,
|
|
"step": 18
|
|
},
|
|
{
|
|
"epoch": 0.08135956108657835,
|
|
"grad_norm": 0.3956957459449768,
|
|
"learning_rate": 6e-05,
|
|
"loss": 0.1025,
|
|
"step": 19
|
|
},
|
|
{
|
|
"epoch": 0.08564164324902984,
|
|
"grad_norm": 0.40594249963760376,
|
|
"learning_rate": 6.333333333333333e-05,
|
|
"loss": 0.0923,
|
|
"step": 20
|
|
},
|
|
{
|
|
"epoch": 0.08992372541148133,
|
|
"grad_norm": 0.4704301059246063,
|
|
"learning_rate": 6.666666666666667e-05,
|
|
"loss": 0.119,
|
|
"step": 21
|
|
},
|
|
{
|
|
"epoch": 0.09420580757393282,
|
|
"grad_norm": 0.44591984152793884,
|
|
"learning_rate": 7e-05,
|
|
"loss": 0.1016,
|
|
"step": 22
|
|
},
|
|
{
|
|
"epoch": 0.09848788973638432,
|
|
"grad_norm": 0.4180421233177185,
|
|
"learning_rate": 7.333333333333333e-05,
|
|
"loss": 0.1043,
|
|
"step": 23
|
|
},
|
|
{
|
|
"epoch": 0.10276997189883581,
|
|
"grad_norm": 0.36205992102622986,
|
|
"learning_rate": 7.666666666666667e-05,
|
|
"loss": 0.0952,
|
|
"step": 24
|
|
},
|
|
{
|
|
"epoch": 0.1070520540612873,
|
|
"grad_norm": 0.328995019197464,
|
|
"learning_rate": 8e-05,
|
|
"loss": 0.0846,
|
|
"step": 25
|
|
},
|
|
{
|
|
"epoch": 0.11133413622373879,
|
|
"grad_norm": 0.3208652138710022,
|
|
"learning_rate": 8.333333333333334e-05,
|
|
"loss": 0.0886,
|
|
"step": 26
|
|
},
|
|
{
|
|
"epoch": 0.11561621838619028,
|
|
"grad_norm": 0.40951845049858093,
|
|
"learning_rate": 8.666666666666667e-05,
|
|
"loss": 0.086,
|
|
"step": 27
|
|
},
|
|
{
|
|
"epoch": 0.11989830054864177,
|
|
"grad_norm": 0.42961806058883667,
|
|
"learning_rate": 9e-05,
|
|
"loss": 0.0922,
|
|
"step": 28
|
|
},
|
|
{
|
|
"epoch": 0.12418038271109327,
|
|
"grad_norm": 0.36131271719932556,
|
|
"learning_rate": 9.333333333333334e-05,
|
|
"loss": 0.0837,
|
|
"step": 29
|
|
},
|
|
{
|
|
"epoch": 0.12846246487354476,
|
|
"grad_norm": 0.3856489360332489,
|
|
"learning_rate": 9.666666666666667e-05,
|
|
"loss": 0.0874,
|
|
"step": 30
|
|
},
|
|
{
|
|
"epoch": 0.13274454703599625,
|
|
"grad_norm": 0.3986302614212036,
|
|
"learning_rate": 0.0001,
|
|
"loss": 0.0846,
|
|
"step": 31
|
|
},
|
|
{
|
|
"epoch": 0.13702662919844774,
|
|
"grad_norm": 0.40041351318359375,
|
|
"learning_rate": 9.950980392156864e-05,
|
|
"loss": 0.0878,
|
|
"step": 32
|
|
},
|
|
{
|
|
"epoch": 0.14130871136089923,
|
|
"grad_norm": 0.45992231369018555,
|
|
"learning_rate": 9.901960784313727e-05,
|
|
"loss": 0.0851,
|
|
"step": 33
|
|
},
|
|
{
|
|
"epoch": 0.14559079352335072,
|
|
"grad_norm": 0.41897663474082947,
|
|
"learning_rate": 9.852941176470589e-05,
|
|
"loss": 0.0835,
|
|
"step": 34
|
|
},
|
|
{
|
|
"epoch": 0.1498728756858022,
|
|
"grad_norm": 0.3854196071624756,
|
|
"learning_rate": 9.80392156862745e-05,
|
|
"loss": 0.0649,
|
|
"step": 35
|
|
},
|
|
{
|
|
"epoch": 0.15415495784825373,
|
|
"grad_norm": 0.3468564748764038,
|
|
"learning_rate": 9.754901960784314e-05,
|
|
"loss": 0.0619,
|
|
"step": 36
|
|
},
|
|
{
|
|
"epoch": 0.15843704001070522,
|
|
"grad_norm": 0.3781249225139618,
|
|
"learning_rate": 9.705882352941177e-05,
|
|
"loss": 0.0723,
|
|
"step": 37
|
|
},
|
|
{
|
|
"epoch": 0.1627191221731567,
|
|
"grad_norm": 0.33684301376342773,
|
|
"learning_rate": 9.65686274509804e-05,
|
|
"loss": 0.0619,
|
|
"step": 38
|
|
},
|
|
{
|
|
"epoch": 0.1670012043356082,
|
|
"grad_norm": 0.319573312997818,
|
|
"learning_rate": 9.607843137254903e-05,
|
|
"loss": 0.0711,
|
|
"step": 39
|
|
},
|
|
{
|
|
"epoch": 0.17128328649805968,
|
|
"grad_norm": 0.4080142378807068,
|
|
"learning_rate": 9.558823529411765e-05,
|
|
"loss": 0.0637,
|
|
"step": 40
|
|
},
|
|
{
|
|
"epoch": 0.17556536866051117,
|
|
"grad_norm": 0.34714841842651367,
|
|
"learning_rate": 9.509803921568627e-05,
|
|
"loss": 0.0666,
|
|
"step": 41
|
|
},
|
|
{
|
|
"epoch": 0.17984745082296266,
|
|
"grad_norm": 0.3296521306037903,
|
|
"learning_rate": 9.46078431372549e-05,
|
|
"loss": 0.0639,
|
|
"step": 42
|
|
},
|
|
{
|
|
"epoch": 0.18412953298541415,
|
|
"grad_norm": 0.3505910038948059,
|
|
"learning_rate": 9.411764705882353e-05,
|
|
"loss": 0.0756,
|
|
"step": 43
|
|
},
|
|
{
|
|
"epoch": 0.18841161514786564,
|
|
"grad_norm": 0.37628045678138733,
|
|
"learning_rate": 9.362745098039217e-05,
|
|
"loss": 0.0605,
|
|
"step": 44
|
|
},
|
|
{
|
|
"epoch": 0.19269369731031713,
|
|
"grad_norm": 0.3756776452064514,
|
|
"learning_rate": 9.313725490196079e-05,
|
|
"loss": 0.0695,
|
|
"step": 45
|
|
},
|
|
{
|
|
"epoch": 0.19697577947276865,
|
|
"grad_norm": 0.3502165675163269,
|
|
"learning_rate": 9.264705882352942e-05,
|
|
"loss": 0.0715,
|
|
"step": 46
|
|
},
|
|
{
|
|
"epoch": 0.20125786163522014,
|
|
"grad_norm": 0.38754549622535706,
|
|
"learning_rate": 9.215686274509804e-05,
|
|
"loss": 0.0619,
|
|
"step": 47
|
|
},
|
|
{
|
|
"epoch": 0.20553994379767163,
|
|
"grad_norm": 0.342479407787323,
|
|
"learning_rate": 9.166666666666667e-05,
|
|
"loss": 0.0697,
|
|
"step": 48
|
|
},
|
|
{
|
|
"epoch": 0.20982202596012312,
|
|
"grad_norm": 0.34569641947746277,
|
|
"learning_rate": 9.11764705882353e-05,
|
|
"loss": 0.0608,
|
|
"step": 49
|
|
},
|
|
{
|
|
"epoch": 0.2141041081225746,
|
|
"grad_norm": 0.3148820996284485,
|
|
"learning_rate": 9.068627450980393e-05,
|
|
"loss": 0.0553,
|
|
"step": 50
|
|
},
|
|
{
|
|
"epoch": 0.2183861902850261,
|
|
"grad_norm": 0.2726571559906006,
|
|
"learning_rate": 9.019607843137255e-05,
|
|
"loss": 0.0525,
|
|
"step": 51
|
|
},
|
|
{
|
|
"epoch": 0.22266827244747758,
|
|
"grad_norm": 0.2933000326156616,
|
|
"learning_rate": 8.970588235294118e-05,
|
|
"loss": 0.0529,
|
|
"step": 52
|
|
},
|
|
{
|
|
"epoch": 0.22695035460992907,
|
|
"grad_norm": 0.32735496759414673,
|
|
"learning_rate": 8.921568627450981e-05,
|
|
"loss": 0.0654,
|
|
"step": 53
|
|
},
|
|
{
|
|
"epoch": 0.23123243677238056,
|
|
"grad_norm": 0.3951069712638855,
|
|
"learning_rate": 8.872549019607843e-05,
|
|
"loss": 0.0672,
|
|
"step": 54
|
|
},
|
|
{
|
|
"epoch": 0.23551451893483205,
|
|
"grad_norm": 0.3004348874092102,
|
|
"learning_rate": 8.823529411764706e-05,
|
|
"loss": 0.0473,
|
|
"step": 55
|
|
},
|
|
{
|
|
"epoch": 0.23979660109728354,
|
|
"grad_norm": 0.3402177691459656,
|
|
"learning_rate": 8.774509803921568e-05,
|
|
"loss": 0.066,
|
|
"step": 56
|
|
},
|
|
{
|
|
"epoch": 0.24407868325973506,
|
|
"grad_norm": 0.33595505356788635,
|
|
"learning_rate": 8.725490196078432e-05,
|
|
"loss": 0.0514,
|
|
"step": 57
|
|
},
|
|
{
|
|
"epoch": 0.24836076542218655,
|
|
"grad_norm": 0.337112158536911,
|
|
"learning_rate": 8.676470588235295e-05,
|
|
"loss": 0.0572,
|
|
"step": 58
|
|
},
|
|
{
|
|
"epoch": 0.252642847584638,
|
|
"grad_norm": 0.33211183547973633,
|
|
"learning_rate": 8.627450980392158e-05,
|
|
"loss": 0.053,
|
|
"step": 59
|
|
},
|
|
{
|
|
"epoch": 0.2569249297470895,
|
|
"grad_norm": 0.42650142312049866,
|
|
"learning_rate": 8.57843137254902e-05,
|
|
"loss": 0.063,
|
|
"step": 60
|
|
},
|
|
{
|
|
"epoch": 0.261207011909541,
|
|
"grad_norm": 0.34502434730529785,
|
|
"learning_rate": 8.529411764705883e-05,
|
|
"loss": 0.0656,
|
|
"step": 61
|
|
},
|
|
{
|
|
"epoch": 0.2654890940719925,
|
|
"grad_norm": 0.3582489490509033,
|
|
"learning_rate": 8.480392156862745e-05,
|
|
"loss": 0.0476,
|
|
"step": 62
|
|
},
|
|
{
|
|
"epoch": 0.269771176234444,
|
|
"grad_norm": 0.27663132548332214,
|
|
"learning_rate": 8.431372549019608e-05,
|
|
"loss": 0.0545,
|
|
"step": 63
|
|
},
|
|
{
|
|
"epoch": 0.2740532583968955,
|
|
"grad_norm": 0.3107115924358368,
|
|
"learning_rate": 8.382352941176471e-05,
|
|
"loss": 0.0412,
|
|
"step": 64
|
|
},
|
|
{
|
|
"epoch": 0.278335340559347,
|
|
"grad_norm": 0.36757537722587585,
|
|
"learning_rate": 8.333333333333334e-05,
|
|
"loss": 0.0468,
|
|
"step": 65
|
|
},
|
|
{
|
|
"epoch": 0.28261742272179846,
|
|
"grad_norm": 0.33179935812950134,
|
|
"learning_rate": 8.284313725490198e-05,
|
|
"loss": 0.0601,
|
|
"step": 66
|
|
},
|
|
{
|
|
"epoch": 0.28689950488425,
|
|
"grad_norm": 0.40680983662605286,
|
|
"learning_rate": 8.23529411764706e-05,
|
|
"loss": 0.0565,
|
|
"step": 67
|
|
},
|
|
{
|
|
"epoch": 0.29118158704670144,
|
|
"grad_norm": 0.3238675594329834,
|
|
"learning_rate": 8.186274509803921e-05,
|
|
"loss": 0.0488,
|
|
"step": 68
|
|
},
|
|
{
|
|
"epoch": 0.29546366920915296,
|
|
"grad_norm": 0.39186710119247437,
|
|
"learning_rate": 8.137254901960785e-05,
|
|
"loss": 0.0505,
|
|
"step": 69
|
|
},
|
|
{
|
|
"epoch": 0.2997457513716044,
|
|
"grad_norm": 0.2939775288105011,
|
|
"learning_rate": 8.088235294117648e-05,
|
|
"loss": 0.0484,
|
|
"step": 70
|
|
},
|
|
{
|
|
"epoch": 0.30402783353405594,
|
|
"grad_norm": 0.37278881669044495,
|
|
"learning_rate": 8.039215686274511e-05,
|
|
"loss": 0.0504,
|
|
"step": 71
|
|
},
|
|
{
|
|
"epoch": 0.30830991569650745,
|
|
"grad_norm": 0.3936941921710968,
|
|
"learning_rate": 7.990196078431373e-05,
|
|
"loss": 0.055,
|
|
"step": 72
|
|
},
|
|
{
|
|
"epoch": 0.3125919978589589,
|
|
"grad_norm": 0.38660097122192383,
|
|
"learning_rate": 7.941176470588235e-05,
|
|
"loss": 0.0821,
|
|
"step": 73
|
|
},
|
|
{
|
|
"epoch": 0.31687408002141043,
|
|
"grad_norm": 0.30945825576782227,
|
|
"learning_rate": 7.892156862745098e-05,
|
|
"loss": 0.0442,
|
|
"step": 74
|
|
},
|
|
{
|
|
"epoch": 0.3211561621838619,
|
|
"grad_norm": 0.31412258744239807,
|
|
"learning_rate": 7.843137254901961e-05,
|
|
"loss": 0.0419,
|
|
"step": 75
|
|
},
|
|
{
|
|
"epoch": 0.3254382443463134,
|
|
"grad_norm": 0.34703952074050903,
|
|
"learning_rate": 7.794117647058824e-05,
|
|
"loss": 0.0575,
|
|
"step": 76
|
|
},
|
|
{
|
|
"epoch": 0.32972032650876487,
|
|
"grad_norm": 0.346296489238739,
|
|
"learning_rate": 7.745098039215687e-05,
|
|
"loss": 0.0603,
|
|
"step": 77
|
|
},
|
|
{
|
|
"epoch": 0.3340024086712164,
|
|
"grad_norm": 0.2502746284008026,
|
|
"learning_rate": 7.696078431372549e-05,
|
|
"loss": 0.0417,
|
|
"step": 78
|
|
},
|
|
{
|
|
"epoch": 0.33828449083366785,
|
|
"grad_norm": 0.3630567193031311,
|
|
"learning_rate": 7.647058823529411e-05,
|
|
"loss": 0.0591,
|
|
"step": 79
|
|
},
|
|
{
|
|
"epoch": 0.34256657299611937,
|
|
"grad_norm": 0.33694136142730713,
|
|
"learning_rate": 7.598039215686274e-05,
|
|
"loss": 0.047,
|
|
"step": 80
|
|
},
|
|
{
|
|
"epoch": 0.34684865515857083,
|
|
"grad_norm": 0.22548595070838928,
|
|
"learning_rate": 7.549019607843137e-05,
|
|
"loss": 0.0448,
|
|
"step": 81
|
|
},
|
|
{
|
|
"epoch": 0.35113073732102235,
|
|
"grad_norm": 0.3071158528327942,
|
|
"learning_rate": 7.500000000000001e-05,
|
|
"loss": 0.0503,
|
|
"step": 82
|
|
},
|
|
{
|
|
"epoch": 0.35541281948347386,
|
|
"grad_norm": 0.3624606132507324,
|
|
"learning_rate": 7.450980392156864e-05,
|
|
"loss": 0.0475,
|
|
"step": 83
|
|
},
|
|
{
|
|
"epoch": 0.3596949016459253,
|
|
"grad_norm": 0.23342928290367126,
|
|
"learning_rate": 7.401960784313726e-05,
|
|
"loss": 0.0337,
|
|
"step": 84
|
|
},
|
|
{
|
|
"epoch": 0.36397698380837684,
|
|
"grad_norm": 0.2770336866378784,
|
|
"learning_rate": 7.352941176470589e-05,
|
|
"loss": 0.0453,
|
|
"step": 85
|
|
},
|
|
{
|
|
"epoch": 0.3682590659708283,
|
|
"grad_norm": 0.27268415689468384,
|
|
"learning_rate": 7.303921568627451e-05,
|
|
"loss": 0.0508,
|
|
"step": 86
|
|
},
|
|
{
|
|
"epoch": 0.3725411481332798,
|
|
"grad_norm": 0.22464269399642944,
|
|
"learning_rate": 7.254901960784314e-05,
|
|
"loss": 0.0348,
|
|
"step": 87
|
|
},
|
|
{
|
|
"epoch": 0.3768232302957313,
|
|
"grad_norm": 0.27807021141052246,
|
|
"learning_rate": 7.205882352941177e-05,
|
|
"loss": 0.0597,
|
|
"step": 88
|
|
},
|
|
{
|
|
"epoch": 0.3811053124581828,
|
|
"grad_norm": 0.30591973662376404,
|
|
"learning_rate": 7.156862745098039e-05,
|
|
"loss": 0.0532,
|
|
"step": 89
|
|
},
|
|
{
|
|
"epoch": 0.38538739462063426,
|
|
"grad_norm": 0.29489806294441223,
|
|
"learning_rate": 7.107843137254902e-05,
|
|
"loss": 0.057,
|
|
"step": 90
|
|
},
|
|
{
|
|
"epoch": 0.3896694767830858,
|
|
"grad_norm": 0.27410319447517395,
|
|
"learning_rate": 7.058823529411765e-05,
|
|
"loss": 0.0529,
|
|
"step": 91
|
|
},
|
|
{
|
|
"epoch": 0.3939515589455373,
|
|
"grad_norm": 0.24769103527069092,
|
|
"learning_rate": 7.009803921568627e-05,
|
|
"loss": 0.0371,
|
|
"step": 92
|
|
},
|
|
{
|
|
"epoch": 0.39823364110798876,
|
|
"grad_norm": 0.30047643184661865,
|
|
"learning_rate": 6.96078431372549e-05,
|
|
"loss": 0.06,
|
|
"step": 93
|
|
},
|
|
{
|
|
"epoch": 0.4025157232704403,
|
|
"grad_norm": 0.2860408425331116,
|
|
"learning_rate": 6.911764705882354e-05,
|
|
"loss": 0.046,
|
|
"step": 94
|
|
},
|
|
{
|
|
"epoch": 0.40679780543289173,
|
|
"grad_norm": 0.2957577407360077,
|
|
"learning_rate": 6.862745098039216e-05,
|
|
"loss": 0.0585,
|
|
"step": 95
|
|
},
|
|
{
|
|
"epoch": 0.41107988759534325,
|
|
"grad_norm": 0.3263351023197174,
|
|
"learning_rate": 6.813725490196079e-05,
|
|
"loss": 0.0538,
|
|
"step": 96
|
|
},
|
|
{
|
|
"epoch": 0.4153619697577947,
|
|
"grad_norm": 0.35550421476364136,
|
|
"learning_rate": 6.764705882352942e-05,
|
|
"loss": 0.0533,
|
|
"step": 97
|
|
},
|
|
{
|
|
"epoch": 0.41964405192024623,
|
|
"grad_norm": 0.2950674295425415,
|
|
"learning_rate": 6.715686274509804e-05,
|
|
"loss": 0.05,
|
|
"step": 98
|
|
},
|
|
{
|
|
"epoch": 0.4239261340826977,
|
|
"grad_norm": 0.2513452172279358,
|
|
"learning_rate": 6.666666666666667e-05,
|
|
"loss": 0.0361,
|
|
"step": 99
|
|
},
|
|
{
|
|
"epoch": 0.4282082162451492,
|
|
"grad_norm": 0.3161787688732147,
|
|
"learning_rate": 6.61764705882353e-05,
|
|
"loss": 0.0483,
|
|
"step": 100
|
|
},
|
|
{
|
|
"epoch": 0.43249029840760067,
|
|
"grad_norm": 0.3478478789329529,
|
|
"learning_rate": 6.568627450980392e-05,
|
|
"loss": 0.0467,
|
|
"step": 101
|
|
},
|
|
{
|
|
"epoch": 0.4367723805700522,
|
|
"grad_norm": 0.33523473143577576,
|
|
"learning_rate": 6.519607843137255e-05,
|
|
"loss": 0.0593,
|
|
"step": 102
|
|
},
|
|
{
|
|
"epoch": 0.4410544627325037,
|
|
"grad_norm": 0.36270254850387573,
|
|
"learning_rate": 6.470588235294118e-05,
|
|
"loss": 0.061,
|
|
"step": 103
|
|
},
|
|
{
|
|
"epoch": 0.44533654489495517,
|
|
"grad_norm": 0.31555601954460144,
|
|
"learning_rate": 6.421568627450982e-05,
|
|
"loss": 0.0404,
|
|
"step": 104
|
|
},
|
|
{
|
|
"epoch": 0.4496186270574067,
|
|
"grad_norm": 0.2953006625175476,
|
|
"learning_rate": 6.372549019607843e-05,
|
|
"loss": 0.0502,
|
|
"step": 105
|
|
},
|
|
{
|
|
"epoch": 0.45390070921985815,
|
|
"grad_norm": 0.30984699726104736,
|
|
"learning_rate": 6.323529411764705e-05,
|
|
"loss": 0.0594,
|
|
"step": 106
|
|
},
|
|
{
|
|
"epoch": 0.45818279138230966,
|
|
"grad_norm": 0.30836424231529236,
|
|
"learning_rate": 6.274509803921569e-05,
|
|
"loss": 0.04,
|
|
"step": 107
|
|
},
|
|
{
|
|
"epoch": 0.4624648735447611,
|
|
"grad_norm": 0.27564123272895813,
|
|
"learning_rate": 6.225490196078432e-05,
|
|
"loss": 0.0372,
|
|
"step": 108
|
|
},
|
|
{
|
|
"epoch": 0.46674695570721264,
|
|
"grad_norm": 0.31601759791374207,
|
|
"learning_rate": 6.176470588235295e-05,
|
|
"loss": 0.0483,
|
|
"step": 109
|
|
},
|
|
{
|
|
"epoch": 0.4710290378696641,
|
|
"grad_norm": 0.40431514382362366,
|
|
"learning_rate": 6.127450980392158e-05,
|
|
"loss": 0.0531,
|
|
"step": 110
|
|
},
|
|
{
|
|
"epoch": 0.4753111200321156,
|
|
"grad_norm": 0.2956823408603668,
|
|
"learning_rate": 6.078431372549019e-05,
|
|
"loss": 0.0393,
|
|
"step": 111
|
|
},
|
|
{
|
|
"epoch": 0.4795932021945671,
|
|
"grad_norm": 0.26746317744255066,
|
|
"learning_rate": 6.0294117647058825e-05,
|
|
"loss": 0.0435,
|
|
"step": 112
|
|
},
|
|
{
|
|
"epoch": 0.4838752843570186,
|
|
"grad_norm": 0.29554516077041626,
|
|
"learning_rate": 5.980392156862745e-05,
|
|
"loss": 0.0419,
|
|
"step": 113
|
|
},
|
|
{
|
|
"epoch": 0.4881573665194701,
|
|
"grad_norm": 0.37722569704055786,
|
|
"learning_rate": 5.931372549019608e-05,
|
|
"loss": 0.0512,
|
|
"step": 114
|
|
},
|
|
{
|
|
"epoch": 0.4924394486819216,
|
|
"grad_norm": 0.2477983981370926,
|
|
"learning_rate": 5.882352941176471e-05,
|
|
"loss": 0.038,
|
|
"step": 115
|
|
},
|
|
{
|
|
"epoch": 0.4967215308443731,
|
|
"grad_norm": 0.22754943370819092,
|
|
"learning_rate": 5.833333333333334e-05,
|
|
"loss": 0.0377,
|
|
"step": 116
|
|
},
|
|
{
|
|
"epoch": 0.5010036130068246,
|
|
"grad_norm": 0.25772324204444885,
|
|
"learning_rate": 5.784313725490197e-05,
|
|
"loss": 0.036,
|
|
"step": 117
|
|
},
|
|
{
|
|
"epoch": 0.505285695169276,
|
|
"grad_norm": 0.2738616168498993,
|
|
"learning_rate": 5.735294117647059e-05,
|
|
"loss": 0.0581,
|
|
"step": 118
|
|
},
|
|
{
|
|
"epoch": 0.5095677773317275,
|
|
"grad_norm": 0.31121647357940674,
|
|
"learning_rate": 5.6862745098039215e-05,
|
|
"loss": 0.0479,
|
|
"step": 119
|
|
},
|
|
{
|
|
"epoch": 0.513849859494179,
|
|
"grad_norm": 0.24819445610046387,
|
|
"learning_rate": 5.637254901960785e-05,
|
|
"loss": 0.0375,
|
|
"step": 120
|
|
},
|
|
{
|
|
"epoch": 0.5181319416566306,
|
|
"grad_norm": 0.3565673232078552,
|
|
"learning_rate": 5.588235294117647e-05,
|
|
"loss": 0.0673,
|
|
"step": 121
|
|
},
|
|
{
|
|
"epoch": 0.522414023819082,
|
|
"grad_norm": 0.2924639880657196,
|
|
"learning_rate": 5.5392156862745104e-05,
|
|
"loss": 0.0439,
|
|
"step": 122
|
|
},
|
|
{
|
|
"epoch": 0.5266961059815335,
|
|
"grad_norm": 0.2511173486709595,
|
|
"learning_rate": 5.490196078431373e-05,
|
|
"loss": 0.043,
|
|
"step": 123
|
|
},
|
|
{
|
|
"epoch": 0.530978188143985,
|
|
"grad_norm": 0.3577595353126526,
|
|
"learning_rate": 5.441176470588235e-05,
|
|
"loss": 0.0573,
|
|
"step": 124
|
|
},
|
|
{
|
|
"epoch": 0.5352602703064365,
|
|
"grad_norm": 0.2690516412258148,
|
|
"learning_rate": 5.392156862745098e-05,
|
|
"loss": 0.0462,
|
|
"step": 125
|
|
},
|
|
{
|
|
"epoch": 0.539542352468888,
|
|
"grad_norm": 0.2612890303134918,
|
|
"learning_rate": 5.343137254901961e-05,
|
|
"loss": 0.0452,
|
|
"step": 126
|
|
},
|
|
{
|
|
"epoch": 0.5438244346313394,
|
|
"grad_norm": 0.27475517988204956,
|
|
"learning_rate": 5.294117647058824e-05,
|
|
"loss": 0.0436,
|
|
"step": 127
|
|
},
|
|
{
|
|
"epoch": 0.548106516793791,
|
|
"grad_norm": 0.33208557963371277,
|
|
"learning_rate": 5.245098039215687e-05,
|
|
"loss": 0.0508,
|
|
"step": 128
|
|
},
|
|
{
|
|
"epoch": 0.5523885989562425,
|
|
"grad_norm": 0.28704285621643066,
|
|
"learning_rate": 5.1960784313725495e-05,
|
|
"loss": 0.0436,
|
|
"step": 129
|
|
},
|
|
{
|
|
"epoch": 0.556670681118694,
|
|
"grad_norm": 0.35237687826156616,
|
|
"learning_rate": 5.147058823529411e-05,
|
|
"loss": 0.0549,
|
|
"step": 130
|
|
},
|
|
{
|
|
"epoch": 0.5609527632811454,
|
|
"grad_norm": 0.290622740983963,
|
|
"learning_rate": 5.0980392156862745e-05,
|
|
"loss": 0.048,
|
|
"step": 131
|
|
},
|
|
{
|
|
"epoch": 0.5652348454435969,
|
|
"grad_norm": 0.250241219997406,
|
|
"learning_rate": 5.049019607843137e-05,
|
|
"loss": 0.0441,
|
|
"step": 132
|
|
},
|
|
{
|
|
"epoch": 0.5695169276060484,
|
|
"grad_norm": 0.3002930283546448,
|
|
"learning_rate": 5e-05,
|
|
"loss": 0.0411,
|
|
"step": 133
|
|
},
|
|
{
|
|
"epoch": 0.5737990097685,
|
|
"grad_norm": 0.30466729402542114,
|
|
"learning_rate": 4.9509803921568634e-05,
|
|
"loss": 0.0452,
|
|
"step": 134
|
|
},
|
|
{
|
|
"epoch": 0.5780810919309515,
|
|
"grad_norm": 0.2513613998889923,
|
|
"learning_rate": 4.901960784313725e-05,
|
|
"loss": 0.0408,
|
|
"step": 135
|
|
},
|
|
{
|
|
"epoch": 0.5823631740934029,
|
|
"grad_norm": 0.30394110083580017,
|
|
"learning_rate": 4.8529411764705885e-05,
|
|
"loss": 0.0451,
|
|
"step": 136
|
|
},
|
|
{
|
|
"epoch": 0.5866452562558544,
|
|
"grad_norm": 0.2374386042356491,
|
|
"learning_rate": 4.803921568627452e-05,
|
|
"loss": 0.0391,
|
|
"step": 137
|
|
},
|
|
{
|
|
"epoch": 0.5909273384183059,
|
|
"grad_norm": 0.3214227259159088,
|
|
"learning_rate": 4.7549019607843135e-05,
|
|
"loss": 0.0428,
|
|
"step": 138
|
|
},
|
|
{
|
|
"epoch": 0.5952094205807574,
|
|
"grad_norm": 0.25820499658584595,
|
|
"learning_rate": 4.705882352941177e-05,
|
|
"loss": 0.0458,
|
|
"step": 139
|
|
},
|
|
{
|
|
"epoch": 0.5994915027432088,
|
|
"grad_norm": 0.21727867424488068,
|
|
"learning_rate": 4.656862745098039e-05,
|
|
"loss": 0.0313,
|
|
"step": 140
|
|
},
|
|
{
|
|
"epoch": 0.6037735849056604,
|
|
"grad_norm": 0.3863449692726135,
|
|
"learning_rate": 4.607843137254902e-05,
|
|
"loss": 0.0578,
|
|
"step": 141
|
|
},
|
|
{
|
|
"epoch": 0.6080556670681119,
|
|
"grad_norm": 0.2634093463420868,
|
|
"learning_rate": 4.558823529411765e-05,
|
|
"loss": 0.0373,
|
|
"step": 142
|
|
},
|
|
{
|
|
"epoch": 0.6123377492305634,
|
|
"grad_norm": 0.2833387553691864,
|
|
"learning_rate": 4.5098039215686275e-05,
|
|
"loss": 0.051,
|
|
"step": 143
|
|
},
|
|
{
|
|
"epoch": 0.6166198313930149,
|
|
"grad_norm": 0.3331435024738312,
|
|
"learning_rate": 4.460784313725491e-05,
|
|
"loss": 0.0639,
|
|
"step": 144
|
|
},
|
|
{
|
|
"epoch": 0.6209019135554663,
|
|
"grad_norm": 0.2673349380493164,
|
|
"learning_rate": 4.411764705882353e-05,
|
|
"loss": 0.0509,
|
|
"step": 145
|
|
},
|
|
{
|
|
"epoch": 0.6251839957179178,
|
|
"grad_norm": 0.3010377585887909,
|
|
"learning_rate": 4.362745098039216e-05,
|
|
"loss": 0.0521,
|
|
"step": 146
|
|
},
|
|
{
|
|
"epoch": 0.6294660778803693,
|
|
"grad_norm": 0.28252658247947693,
|
|
"learning_rate": 4.313725490196079e-05,
|
|
"loss": 0.0455,
|
|
"step": 147
|
|
},
|
|
{
|
|
"epoch": 0.6337481600428209,
|
|
"grad_norm": 0.28520438075065613,
|
|
"learning_rate": 4.2647058823529415e-05,
|
|
"loss": 0.0551,
|
|
"step": 148
|
|
},
|
|
{
|
|
"epoch": 0.6380302422052723,
|
|
"grad_norm": 0.24448159337043762,
|
|
"learning_rate": 4.215686274509804e-05,
|
|
"loss": 0.0362,
|
|
"step": 149
|
|
},
|
|
{
|
|
"epoch": 0.6423123243677238,
|
|
"grad_norm": 0.2240142524242401,
|
|
"learning_rate": 4.166666666666667e-05,
|
|
"loss": 0.0377,
|
|
"step": 150
|
|
},
|
|
{
|
|
"epoch": 0.6465944065301753,
|
|
"grad_norm": 0.2896118760108948,
|
|
"learning_rate": 4.11764705882353e-05,
|
|
"loss": 0.0378,
|
|
"step": 151
|
|
},
|
|
{
|
|
"epoch": 0.6508764886926268,
|
|
"grad_norm": 0.2663087844848633,
|
|
"learning_rate": 4.068627450980392e-05,
|
|
"loss": 0.0407,
|
|
"step": 152
|
|
},
|
|
{
|
|
"epoch": 0.6551585708550783,
|
|
"grad_norm": 0.3086792230606079,
|
|
"learning_rate": 4.0196078431372555e-05,
|
|
"loss": 0.0323,
|
|
"step": 153
|
|
},
|
|
{
|
|
"epoch": 0.6594406530175297,
|
|
"grad_norm": 0.2776995897293091,
|
|
"learning_rate": 3.970588235294117e-05,
|
|
"loss": 0.0437,
|
|
"step": 154
|
|
},
|
|
{
|
|
"epoch": 0.6637227351799813,
|
|
"grad_norm": 0.3368375897407532,
|
|
"learning_rate": 3.9215686274509805e-05,
|
|
"loss": 0.0525,
|
|
"step": 155
|
|
},
|
|
{
|
|
"epoch": 0.6680048173424328,
|
|
"grad_norm": 0.25399115681648254,
|
|
"learning_rate": 3.872549019607844e-05,
|
|
"loss": 0.0352,
|
|
"step": 156
|
|
},
|
|
{
|
|
"epoch": 0.6722868995048843,
|
|
"grad_norm": 0.3201920986175537,
|
|
"learning_rate": 3.8235294117647055e-05,
|
|
"loss": 0.046,
|
|
"step": 157
|
|
},
|
|
{
|
|
"epoch": 0.6765689816673357,
|
|
"grad_norm": 0.36324596405029297,
|
|
"learning_rate": 3.774509803921569e-05,
|
|
"loss": 0.0429,
|
|
"step": 158
|
|
},
|
|
{
|
|
"epoch": 0.6808510638297872,
|
|
"grad_norm": 0.27535951137542725,
|
|
"learning_rate": 3.725490196078432e-05,
|
|
"loss": 0.036,
|
|
"step": 159
|
|
},
|
|
{
|
|
"epoch": 0.6851331459922387,
|
|
"grad_norm": 0.2501991391181946,
|
|
"learning_rate": 3.6764705882352945e-05,
|
|
"loss": 0.0333,
|
|
"step": 160
|
|
},
|
|
{
|
|
"epoch": 0.6894152281546903,
|
|
"grad_norm": 0.26270344853401184,
|
|
"learning_rate": 3.627450980392157e-05,
|
|
"loss": 0.0359,
|
|
"step": 161
|
|
},
|
|
{
|
|
"epoch": 0.6936973103171417,
|
|
"grad_norm": 0.2966891825199127,
|
|
"learning_rate": 3.5784313725490195e-05,
|
|
"loss": 0.0448,
|
|
"step": 162
|
|
},
|
|
{
|
|
"epoch": 0.6979793924795932,
|
|
"grad_norm": 0.24465243518352509,
|
|
"learning_rate": 3.529411764705883e-05,
|
|
"loss": 0.0488,
|
|
"step": 163
|
|
},
|
|
{
|
|
"epoch": 0.7022614746420447,
|
|
"grad_norm": 0.2697572410106659,
|
|
"learning_rate": 3.480392156862745e-05,
|
|
"loss": 0.0295,
|
|
"step": 164
|
|
},
|
|
{
|
|
"epoch": 0.7065435568044962,
|
|
"grad_norm": 0.27013951539993286,
|
|
"learning_rate": 3.431372549019608e-05,
|
|
"loss": 0.0395,
|
|
"step": 165
|
|
},
|
|
{
|
|
"epoch": 0.7108256389669477,
|
|
"grad_norm": 0.3045112192630768,
|
|
"learning_rate": 3.382352941176471e-05,
|
|
"loss": 0.0411,
|
|
"step": 166
|
|
},
|
|
{
|
|
"epoch": 0.7151077211293991,
|
|
"grad_norm": 0.24719510972499847,
|
|
"learning_rate": 3.3333333333333335e-05,
|
|
"loss": 0.0354,
|
|
"step": 167
|
|
},
|
|
{
|
|
"epoch": 0.7193898032918506,
|
|
"grad_norm": 0.1892707198858261,
|
|
"learning_rate": 3.284313725490196e-05,
|
|
"loss": 0.0234,
|
|
"step": 168
|
|
},
|
|
{
|
|
"epoch": 0.7236718854543022,
|
|
"grad_norm": 0.30045631527900696,
|
|
"learning_rate": 3.235294117647059e-05,
|
|
"loss": 0.0447,
|
|
"step": 169
|
|
},
|
|
{
|
|
"epoch": 0.7279539676167537,
|
|
"grad_norm": 0.309243768453598,
|
|
"learning_rate": 3.186274509803922e-05,
|
|
"loss": 0.0404,
|
|
"step": 170
|
|
},
|
|
{
|
|
"epoch": 0.7322360497792051,
|
|
"grad_norm": 0.24024784564971924,
|
|
"learning_rate": 3.137254901960784e-05,
|
|
"loss": 0.0395,
|
|
"step": 171
|
|
},
|
|
{
|
|
"epoch": 0.7365181319416566,
|
|
"grad_norm": 0.2552880048751831,
|
|
"learning_rate": 3.0882352941176475e-05,
|
|
"loss": 0.0359,
|
|
"step": 172
|
|
},
|
|
{
|
|
"epoch": 0.7408002141041081,
|
|
"grad_norm": 0.33713486790657043,
|
|
"learning_rate": 3.0392156862745097e-05,
|
|
"loss": 0.0614,
|
|
"step": 173
|
|
},
|
|
{
|
|
"epoch": 0.7450822962665596,
|
|
"grad_norm": 0.2314293533563614,
|
|
"learning_rate": 2.9901960784313725e-05,
|
|
"loss": 0.0348,
|
|
"step": 174
|
|
},
|
|
{
|
|
"epoch": 0.7493643784290112,
|
|
"grad_norm": 0.23023654520511627,
|
|
"learning_rate": 2.9411764705882354e-05,
|
|
"loss": 0.0419,
|
|
"step": 175
|
|
},
|
|
{
|
|
"epoch": 0.7536464605914626,
|
|
"grad_norm": 0.2514519691467285,
|
|
"learning_rate": 2.8921568627450986e-05,
|
|
"loss": 0.0388,
|
|
"step": 176
|
|
},
|
|
{
|
|
"epoch": 0.7579285427539141,
|
|
"grad_norm": 0.2647082209587097,
|
|
"learning_rate": 2.8431372549019608e-05,
|
|
"loss": 0.0463,
|
|
"step": 177
|
|
},
|
|
{
|
|
"epoch": 0.7622106249163656,
|
|
"grad_norm": 0.29537469148635864,
|
|
"learning_rate": 2.7941176470588236e-05,
|
|
"loss": 0.0352,
|
|
"step": 178
|
|
},
|
|
{
|
|
"epoch": 0.7664927070788171,
|
|
"grad_norm": 0.2043161392211914,
|
|
"learning_rate": 2.7450980392156865e-05,
|
|
"loss": 0.0387,
|
|
"step": 179
|
|
},
|
|
{
|
|
"epoch": 0.7707747892412685,
|
|
"grad_norm": 0.305243581533432,
|
|
"learning_rate": 2.696078431372549e-05,
|
|
"loss": 0.0346,
|
|
"step": 180
|
|
},
|
|
{
|
|
"epoch": 0.77505687140372,
|
|
"grad_norm": 0.29974740743637085,
|
|
"learning_rate": 2.647058823529412e-05,
|
|
"loss": 0.0513,
|
|
"step": 181
|
|
},
|
|
{
|
|
"epoch": 0.7793389535661716,
|
|
"grad_norm": 0.28913021087646484,
|
|
"learning_rate": 2.5980392156862747e-05,
|
|
"loss": 0.0341,
|
|
"step": 182
|
|
},
|
|
{
|
|
"epoch": 0.7836210357286231,
|
|
"grad_norm": 0.32380354404449463,
|
|
"learning_rate": 2.5490196078431373e-05,
|
|
"loss": 0.0614,
|
|
"step": 183
|
|
},
|
|
{
|
|
"epoch": 0.7879031178910746,
|
|
"grad_norm": 0.2799640893936157,
|
|
"learning_rate": 2.5e-05,
|
|
"loss": 0.0515,
|
|
"step": 184
|
|
},
|
|
{
|
|
"epoch": 0.792185200053526,
|
|
"grad_norm": 0.28914907574653625,
|
|
"learning_rate": 2.4509803921568626e-05,
|
|
"loss": 0.0393,
|
|
"step": 185
|
|
},
|
|
{
|
|
"epoch": 0.7964672822159775,
|
|
"grad_norm": 0.26201391220092773,
|
|
"learning_rate": 2.401960784313726e-05,
|
|
"loss": 0.0431,
|
|
"step": 186
|
|
},
|
|
{
|
|
"epoch": 0.800749364378429,
|
|
"grad_norm": 0.29834750294685364,
|
|
"learning_rate": 2.3529411764705884e-05,
|
|
"loss": 0.039,
|
|
"step": 187
|
|
},
|
|
{
|
|
"epoch": 0.8050314465408805,
|
|
"grad_norm": 0.28359827399253845,
|
|
"learning_rate": 2.303921568627451e-05,
|
|
"loss": 0.0474,
|
|
"step": 188
|
|
},
|
|
{
|
|
"epoch": 0.809313528703332,
|
|
"grad_norm": 0.21148712933063507,
|
|
"learning_rate": 2.2549019607843138e-05,
|
|
"loss": 0.0355,
|
|
"step": 189
|
|
},
|
|
{
|
|
"epoch": 0.8135956108657835,
|
|
"grad_norm": 0.30203911662101746,
|
|
"learning_rate": 2.2058823529411766e-05,
|
|
"loss": 0.0423,
|
|
"step": 190
|
|
},
|
|
{
|
|
"epoch": 0.817877693028235,
|
|
"grad_norm": 0.32118844985961914,
|
|
"learning_rate": 2.1568627450980395e-05,
|
|
"loss": 0.0396,
|
|
"step": 191
|
|
},
|
|
{
|
|
"epoch": 0.8221597751906865,
|
|
"grad_norm": 0.2358664572238922,
|
|
"learning_rate": 2.107843137254902e-05,
|
|
"loss": 0.0344,
|
|
"step": 192
|
|
},
|
|
{
|
|
"epoch": 0.8264418573531379,
|
|
"grad_norm": 0.3052966892719269,
|
|
"learning_rate": 2.058823529411765e-05,
|
|
"loss": 0.0459,
|
|
"step": 193
|
|
},
|
|
{
|
|
"epoch": 0.8307239395155894,
|
|
"grad_norm": 0.2716437578201294,
|
|
"learning_rate": 2.0098039215686277e-05,
|
|
"loss": 0.0511,
|
|
"step": 194
|
|
},
|
|
{
|
|
"epoch": 0.8350060216780409,
|
|
"grad_norm": 0.3086344301700592,
|
|
"learning_rate": 1.9607843137254903e-05,
|
|
"loss": 0.0477,
|
|
"step": 195
|
|
},
|
|
{
|
|
"epoch": 0.8392881038404925,
|
|
"grad_norm": 0.26460692286491394,
|
|
"learning_rate": 1.9117647058823528e-05,
|
|
"loss": 0.0355,
|
|
"step": 196
|
|
},
|
|
{
|
|
"epoch": 0.843570186002944,
|
|
"grad_norm": 0.2656954526901245,
|
|
"learning_rate": 1.862745098039216e-05,
|
|
"loss": 0.0504,
|
|
"step": 197
|
|
},
|
|
{
|
|
"epoch": 0.8478522681653954,
|
|
"grad_norm": 0.2059815526008606,
|
|
"learning_rate": 1.8137254901960785e-05,
|
|
"loss": 0.0404,
|
|
"step": 198
|
|
},
|
|
{
|
|
"epoch": 0.8521343503278469,
|
|
"grad_norm": 0.28695958852767944,
|
|
"learning_rate": 1.7647058823529414e-05,
|
|
"loss": 0.0502,
|
|
"step": 199
|
|
},
|
|
{
|
|
"epoch": 0.8564164324902984,
|
|
"grad_norm": 0.278156042098999,
|
|
"learning_rate": 1.715686274509804e-05,
|
|
"loss": 0.0325,
|
|
"step": 200
|
|
},
|
|
{
|
|
"epoch": 0.8606985146527499,
|
|
"grad_norm": 0.33728834986686707,
|
|
"learning_rate": 1.6666666666666667e-05,
|
|
"loss": 0.0468,
|
|
"step": 201
|
|
},
|
|
{
|
|
"epoch": 0.8649805968152013,
|
|
"grad_norm": 0.2665461003780365,
|
|
"learning_rate": 1.6176470588235296e-05,
|
|
"loss": 0.041,
|
|
"step": 202
|
|
},
|
|
{
|
|
"epoch": 0.8692626789776529,
|
|
"grad_norm": 0.2768334150314331,
|
|
"learning_rate": 1.568627450980392e-05,
|
|
"loss": 0.0477,
|
|
"step": 203
|
|
},
|
|
{
|
|
"epoch": 0.8735447611401044,
|
|
"grad_norm": 0.2847307324409485,
|
|
"learning_rate": 1.5196078431372548e-05,
|
|
"loss": 0.0368,
|
|
"step": 204
|
|
},
|
|
{
|
|
"epoch": 0.8778268433025559,
|
|
"grad_norm": 0.26865965127944946,
|
|
"learning_rate": 1.4705882352941177e-05,
|
|
"loss": 0.0321,
|
|
"step": 205
|
|
},
|
|
{
|
|
"epoch": 0.8821089254650074,
|
|
"grad_norm": 0.23639728128910065,
|
|
"learning_rate": 1.4215686274509804e-05,
|
|
"loss": 0.0267,
|
|
"step": 206
|
|
},
|
|
{
|
|
"epoch": 0.8863910076274588,
|
|
"grad_norm": 0.22090226411819458,
|
|
"learning_rate": 1.3725490196078432e-05,
|
|
"loss": 0.0283,
|
|
"step": 207
|
|
},
|
|
{
|
|
"epoch": 0.8906730897899103,
|
|
"grad_norm": 0.24385924637317657,
|
|
"learning_rate": 1.323529411764706e-05,
|
|
"loss": 0.0486,
|
|
"step": 208
|
|
},
|
|
{
|
|
"epoch": 0.8949551719523618,
|
|
"grad_norm": 0.267991304397583,
|
|
"learning_rate": 1.2745098039215686e-05,
|
|
"loss": 0.049,
|
|
"step": 209
|
|
},
|
|
{
|
|
"epoch": 0.8992372541148134,
|
|
"grad_norm": 0.2858865261077881,
|
|
"learning_rate": 1.2254901960784313e-05,
|
|
"loss": 0.0416,
|
|
"step": 210
|
|
},
|
|
{
|
|
"epoch": 0.9035193362772648,
|
|
"grad_norm": 0.30461645126342773,
|
|
"learning_rate": 1.1764705882352942e-05,
|
|
"loss": 0.0459,
|
|
"step": 211
|
|
},
|
|
{
|
|
"epoch": 0.9078014184397163,
|
|
"grad_norm": 0.3148707449436188,
|
|
"learning_rate": 1.1274509803921569e-05,
|
|
"loss": 0.0448,
|
|
"step": 212
|
|
},
|
|
{
|
|
"epoch": 0.9120835006021678,
|
|
"grad_norm": 0.29912617802619934,
|
|
"learning_rate": 1.0784313725490197e-05,
|
|
"loss": 0.0391,
|
|
"step": 213
|
|
},
|
|
{
|
|
"epoch": 0.9163655827646193,
|
|
"grad_norm": 0.30762436985969543,
|
|
"learning_rate": 1.0294117647058824e-05,
|
|
"loss": 0.0475,
|
|
"step": 214
|
|
},
|
|
{
|
|
"epoch": 0.9206476649270708,
|
|
"grad_norm": 0.27570030093193054,
|
|
"learning_rate": 9.803921568627451e-06,
|
|
"loss": 0.0371,
|
|
"step": 215
|
|
},
|
|
{
|
|
"epoch": 0.9249297470895222,
|
|
"grad_norm": 0.26668745279312134,
|
|
"learning_rate": 9.31372549019608e-06,
|
|
"loss": 0.0532,
|
|
"step": 216
|
|
},
|
|
{
|
|
"epoch": 0.9292118292519738,
|
|
"grad_norm": 0.20003274083137512,
|
|
"learning_rate": 8.823529411764707e-06,
|
|
"loss": 0.0328,
|
|
"step": 217
|
|
},
|
|
{
|
|
"epoch": 0.9334939114144253,
|
|
"grad_norm": 0.23162154853343964,
|
|
"learning_rate": 8.333333333333334e-06,
|
|
"loss": 0.0348,
|
|
"step": 218
|
|
},
|
|
{
|
|
"epoch": 0.9377759935768768,
|
|
"grad_norm": 0.28092819452285767,
|
|
"learning_rate": 7.84313725490196e-06,
|
|
"loss": 0.041,
|
|
"step": 219
|
|
},
|
|
{
|
|
"epoch": 0.9420580757393282,
|
|
"grad_norm": 0.20697717368602753,
|
|
"learning_rate": 7.3529411764705884e-06,
|
|
"loss": 0.0344,
|
|
"step": 220
|
|
},
|
|
{
|
|
"epoch": 0.9463401579017797,
|
|
"grad_norm": 0.2537749111652374,
|
|
"learning_rate": 6.862745098039216e-06,
|
|
"loss": 0.0364,
|
|
"step": 221
|
|
},
|
|
{
|
|
"epoch": 0.9506222400642312,
|
|
"grad_norm": 0.23425333201885223,
|
|
"learning_rate": 6.372549019607843e-06,
|
|
"loss": 0.0391,
|
|
"step": 222
|
|
},
|
|
{
|
|
"epoch": 0.9549043222266828,
|
|
"grad_norm": 0.25059816241264343,
|
|
"learning_rate": 5.882352941176471e-06,
|
|
"loss": 0.0506,
|
|
"step": 223
|
|
},
|
|
{
|
|
"epoch": 0.9591864043891342,
|
|
"grad_norm": 0.2436205893754959,
|
|
"learning_rate": 5.392156862745099e-06,
|
|
"loss": 0.0418,
|
|
"step": 224
|
|
},
|
|
{
|
|
"epoch": 0.9634684865515857,
|
|
"grad_norm": 0.255195289850235,
|
|
"learning_rate": 4.901960784313726e-06,
|
|
"loss": 0.0412,
|
|
"step": 225
|
|
},
|
|
{
|
|
"epoch": 0.9677505687140372,
|
|
"grad_norm": 0.27884790301322937,
|
|
"learning_rate": 4.411764705882353e-06,
|
|
"loss": 0.0497,
|
|
"step": 226
|
|
},
|
|
{
|
|
"epoch": 0.9720326508764887,
|
|
"grad_norm": 0.25363633036613464,
|
|
"learning_rate": 3.92156862745098e-06,
|
|
"loss": 0.0408,
|
|
"step": 227
|
|
},
|
|
{
|
|
"epoch": 0.9763147330389402,
|
|
"grad_norm": 0.30385518074035645,
|
|
"learning_rate": 3.431372549019608e-06,
|
|
"loss": 0.0552,
|
|
"step": 228
|
|
},
|
|
{
|
|
"epoch": 0.9805968152013916,
|
|
"grad_norm": 0.27926793694496155,
|
|
"learning_rate": 2.9411764705882355e-06,
|
|
"loss": 0.036,
|
|
"step": 229
|
|
},
|
|
{
|
|
"epoch": 0.9848788973638432,
|
|
"grad_norm": 0.236245259642601,
|
|
"learning_rate": 2.450980392156863e-06,
|
|
"loss": 0.029,
|
|
"step": 230
|
|
},
|
|
{
|
|
"epoch": 0.9891609795262947,
|
|
"grad_norm": 0.25143948197364807,
|
|
"learning_rate": 1.96078431372549e-06,
|
|
"loss": 0.0303,
|
|
"step": 231
|
|
},
|
|
{
|
|
"epoch": 0.9934430616887462,
|
|
"grad_norm": 0.29792022705078125,
|
|
"learning_rate": 1.4705882352941177e-06,
|
|
"loss": 0.0498,
|
|
"step": 232
|
|
},
|
|
{
|
|
"epoch": 0.9977251438511976,
|
|
"grad_norm": 0.32305002212524414,
|
|
"learning_rate": 9.80392156862745e-07,
|
|
"loss": 0.0375,
|
|
"step": 233
|
|
},
|
|
{
|
|
"epoch": 1.0,
|
|
"grad_norm": 0.4602526128292084,
|
|
"learning_rate": 4.901960784313725e-07,
|
|
"loss": 0.0426,
|
|
"step": 234
|
|
}
|
|
],
|
|
"logging_steps": 1,
|
|
"max_steps": 234,
|
|
"num_input_tokens_seen": 0,
|
|
"num_train_epochs": 1,
|
|
"save_steps": 5000,
|
|
"stateful_callbacks": {
|
|
"TrainerControl": {
|
|
"args": {
|
|
"should_epoch_stop": false,
|
|
"should_evaluate": false,
|
|
"should_log": false,
|
|
"should_save": true,
|
|
"should_training_stop": true
|
|
},
|
|
"attributes": {}
|
|
}
|
|
},
|
|
"total_flos": 2.927681385546916e+17,
|
|
"train_batch_size": 1,
|
|
"trial_name": null,
|
|
"trial_params": null
|
|
}
|