Files
ModelHub XC 22294efd80 初始化项目,由ModelHub XC社区提供模型
Model: longtermrisk/Qwen2.5-32B-Instruct-ftjob-b0fafb674e38
Source: Original Platform
2026-04-10 20:45:21 +08:00

1673 lines
40 KiB
JSON

{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 234,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.004282082162451492,
"grad_norm": 2.3133654594421387,
"learning_rate": 0.0,
"loss": 0.9749,
"step": 1
},
{
"epoch": 0.008564164324902984,
"grad_norm": 1.8335061073303223,
"learning_rate": 3.3333333333333333e-06,
"loss": 0.7648,
"step": 2
},
{
"epoch": 0.012846246487354477,
"grad_norm": 2.2353498935699463,
"learning_rate": 6.666666666666667e-06,
"loss": 0.9076,
"step": 3
},
{
"epoch": 0.017128328649805968,
"grad_norm": 2.1200685501098633,
"learning_rate": 1e-05,
"loss": 0.8221,
"step": 4
},
{
"epoch": 0.02141041081225746,
"grad_norm": 2.185300350189209,
"learning_rate": 1.3333333333333333e-05,
"loss": 0.8972,
"step": 5
},
{
"epoch": 0.025692492974708953,
"grad_norm": 2.383875608444214,
"learning_rate": 1.6666666666666667e-05,
"loss": 0.8918,
"step": 6
},
{
"epoch": 0.029974575137160443,
"grad_norm": 2.272465467453003,
"learning_rate": 2e-05,
"loss": 0.803,
"step": 7
},
{
"epoch": 0.034256657299611935,
"grad_norm": 1.9353933334350586,
"learning_rate": 2.3333333333333336e-05,
"loss": 0.5844,
"step": 8
},
{
"epoch": 0.03853873946206343,
"grad_norm": 1.7865619659423828,
"learning_rate": 2.6666666666666667e-05,
"loss": 0.5373,
"step": 9
},
{
"epoch": 0.04282082162451492,
"grad_norm": 1.5210086107254028,
"learning_rate": 3e-05,
"loss": 0.4767,
"step": 10
},
{
"epoch": 0.04710290378696641,
"grad_norm": 1.1693248748779297,
"learning_rate": 3.3333333333333335e-05,
"loss": 0.3285,
"step": 11
},
{
"epoch": 0.051384985949417906,
"grad_norm": 0.8487532138824463,
"learning_rate": 3.6666666666666666e-05,
"loss": 0.2321,
"step": 12
},
{
"epoch": 0.055667068111869396,
"grad_norm": 0.8226314187049866,
"learning_rate": 4e-05,
"loss": 0.2117,
"step": 13
},
{
"epoch": 0.059949150274320885,
"grad_norm": 0.7278441190719604,
"learning_rate": 4.3333333333333334e-05,
"loss": 0.1951,
"step": 14
},
{
"epoch": 0.06423123243677238,
"grad_norm": 0.6208946704864502,
"learning_rate": 4.666666666666667e-05,
"loss": 0.1739,
"step": 15
},
{
"epoch": 0.06851331459922387,
"grad_norm": 0.44936591386795044,
"learning_rate": 5e-05,
"loss": 0.1144,
"step": 16
},
{
"epoch": 0.07279539676167536,
"grad_norm": 0.42960935831069946,
"learning_rate": 5.333333333333333e-05,
"loss": 0.1089,
"step": 17
},
{
"epoch": 0.07707747892412686,
"grad_norm": 0.4222032427787781,
"learning_rate": 5.666666666666667e-05,
"loss": 0.1103,
"step": 18
},
{
"epoch": 0.08135956108657835,
"grad_norm": 0.3956957459449768,
"learning_rate": 6e-05,
"loss": 0.1025,
"step": 19
},
{
"epoch": 0.08564164324902984,
"grad_norm": 0.40594249963760376,
"learning_rate": 6.333333333333333e-05,
"loss": 0.0923,
"step": 20
},
{
"epoch": 0.08992372541148133,
"grad_norm": 0.4704301059246063,
"learning_rate": 6.666666666666667e-05,
"loss": 0.119,
"step": 21
},
{
"epoch": 0.09420580757393282,
"grad_norm": 0.44591984152793884,
"learning_rate": 7e-05,
"loss": 0.1016,
"step": 22
},
{
"epoch": 0.09848788973638432,
"grad_norm": 0.4180421233177185,
"learning_rate": 7.333333333333333e-05,
"loss": 0.1043,
"step": 23
},
{
"epoch": 0.10276997189883581,
"grad_norm": 0.36205992102622986,
"learning_rate": 7.666666666666667e-05,
"loss": 0.0952,
"step": 24
},
{
"epoch": 0.1070520540612873,
"grad_norm": 0.328995019197464,
"learning_rate": 8e-05,
"loss": 0.0846,
"step": 25
},
{
"epoch": 0.11133413622373879,
"grad_norm": 0.3208652138710022,
"learning_rate": 8.333333333333334e-05,
"loss": 0.0886,
"step": 26
},
{
"epoch": 0.11561621838619028,
"grad_norm": 0.40951845049858093,
"learning_rate": 8.666666666666667e-05,
"loss": 0.086,
"step": 27
},
{
"epoch": 0.11989830054864177,
"grad_norm": 0.42961806058883667,
"learning_rate": 9e-05,
"loss": 0.0922,
"step": 28
},
{
"epoch": 0.12418038271109327,
"grad_norm": 0.36131271719932556,
"learning_rate": 9.333333333333334e-05,
"loss": 0.0837,
"step": 29
},
{
"epoch": 0.12846246487354476,
"grad_norm": 0.3856489360332489,
"learning_rate": 9.666666666666667e-05,
"loss": 0.0874,
"step": 30
},
{
"epoch": 0.13274454703599625,
"grad_norm": 0.3986302614212036,
"learning_rate": 0.0001,
"loss": 0.0846,
"step": 31
},
{
"epoch": 0.13702662919844774,
"grad_norm": 0.40041351318359375,
"learning_rate": 9.950980392156864e-05,
"loss": 0.0878,
"step": 32
},
{
"epoch": 0.14130871136089923,
"grad_norm": 0.45992231369018555,
"learning_rate": 9.901960784313727e-05,
"loss": 0.0851,
"step": 33
},
{
"epoch": 0.14559079352335072,
"grad_norm": 0.41897663474082947,
"learning_rate": 9.852941176470589e-05,
"loss": 0.0835,
"step": 34
},
{
"epoch": 0.1498728756858022,
"grad_norm": 0.3854196071624756,
"learning_rate": 9.80392156862745e-05,
"loss": 0.0649,
"step": 35
},
{
"epoch": 0.15415495784825373,
"grad_norm": 0.3468564748764038,
"learning_rate": 9.754901960784314e-05,
"loss": 0.0619,
"step": 36
},
{
"epoch": 0.15843704001070522,
"grad_norm": 0.3781249225139618,
"learning_rate": 9.705882352941177e-05,
"loss": 0.0723,
"step": 37
},
{
"epoch": 0.1627191221731567,
"grad_norm": 0.33684301376342773,
"learning_rate": 9.65686274509804e-05,
"loss": 0.0619,
"step": 38
},
{
"epoch": 0.1670012043356082,
"grad_norm": 0.319573312997818,
"learning_rate": 9.607843137254903e-05,
"loss": 0.0711,
"step": 39
},
{
"epoch": 0.17128328649805968,
"grad_norm": 0.4080142378807068,
"learning_rate": 9.558823529411765e-05,
"loss": 0.0637,
"step": 40
},
{
"epoch": 0.17556536866051117,
"grad_norm": 0.34714841842651367,
"learning_rate": 9.509803921568627e-05,
"loss": 0.0666,
"step": 41
},
{
"epoch": 0.17984745082296266,
"grad_norm": 0.3296521306037903,
"learning_rate": 9.46078431372549e-05,
"loss": 0.0639,
"step": 42
},
{
"epoch": 0.18412953298541415,
"grad_norm": 0.3505910038948059,
"learning_rate": 9.411764705882353e-05,
"loss": 0.0756,
"step": 43
},
{
"epoch": 0.18841161514786564,
"grad_norm": 0.37628045678138733,
"learning_rate": 9.362745098039217e-05,
"loss": 0.0605,
"step": 44
},
{
"epoch": 0.19269369731031713,
"grad_norm": 0.3756776452064514,
"learning_rate": 9.313725490196079e-05,
"loss": 0.0695,
"step": 45
},
{
"epoch": 0.19697577947276865,
"grad_norm": 0.3502165675163269,
"learning_rate": 9.264705882352942e-05,
"loss": 0.0715,
"step": 46
},
{
"epoch": 0.20125786163522014,
"grad_norm": 0.38754549622535706,
"learning_rate": 9.215686274509804e-05,
"loss": 0.0619,
"step": 47
},
{
"epoch": 0.20553994379767163,
"grad_norm": 0.342479407787323,
"learning_rate": 9.166666666666667e-05,
"loss": 0.0697,
"step": 48
},
{
"epoch": 0.20982202596012312,
"grad_norm": 0.34569641947746277,
"learning_rate": 9.11764705882353e-05,
"loss": 0.0608,
"step": 49
},
{
"epoch": 0.2141041081225746,
"grad_norm": 0.3148820996284485,
"learning_rate": 9.068627450980393e-05,
"loss": 0.0553,
"step": 50
},
{
"epoch": 0.2183861902850261,
"grad_norm": 0.2726571559906006,
"learning_rate": 9.019607843137255e-05,
"loss": 0.0525,
"step": 51
},
{
"epoch": 0.22266827244747758,
"grad_norm": 0.2933000326156616,
"learning_rate": 8.970588235294118e-05,
"loss": 0.0529,
"step": 52
},
{
"epoch": 0.22695035460992907,
"grad_norm": 0.32735496759414673,
"learning_rate": 8.921568627450981e-05,
"loss": 0.0654,
"step": 53
},
{
"epoch": 0.23123243677238056,
"grad_norm": 0.3951069712638855,
"learning_rate": 8.872549019607843e-05,
"loss": 0.0672,
"step": 54
},
{
"epoch": 0.23551451893483205,
"grad_norm": 0.3004348874092102,
"learning_rate": 8.823529411764706e-05,
"loss": 0.0473,
"step": 55
},
{
"epoch": 0.23979660109728354,
"grad_norm": 0.3402177691459656,
"learning_rate": 8.774509803921568e-05,
"loss": 0.066,
"step": 56
},
{
"epoch": 0.24407868325973506,
"grad_norm": 0.33595505356788635,
"learning_rate": 8.725490196078432e-05,
"loss": 0.0514,
"step": 57
},
{
"epoch": 0.24836076542218655,
"grad_norm": 0.337112158536911,
"learning_rate": 8.676470588235295e-05,
"loss": 0.0572,
"step": 58
},
{
"epoch": 0.252642847584638,
"grad_norm": 0.33211183547973633,
"learning_rate": 8.627450980392158e-05,
"loss": 0.053,
"step": 59
},
{
"epoch": 0.2569249297470895,
"grad_norm": 0.42650142312049866,
"learning_rate": 8.57843137254902e-05,
"loss": 0.063,
"step": 60
},
{
"epoch": 0.261207011909541,
"grad_norm": 0.34502434730529785,
"learning_rate": 8.529411764705883e-05,
"loss": 0.0656,
"step": 61
},
{
"epoch": 0.2654890940719925,
"grad_norm": 0.3582489490509033,
"learning_rate": 8.480392156862745e-05,
"loss": 0.0476,
"step": 62
},
{
"epoch": 0.269771176234444,
"grad_norm": 0.27663132548332214,
"learning_rate": 8.431372549019608e-05,
"loss": 0.0545,
"step": 63
},
{
"epoch": 0.2740532583968955,
"grad_norm": 0.3107115924358368,
"learning_rate": 8.382352941176471e-05,
"loss": 0.0412,
"step": 64
},
{
"epoch": 0.278335340559347,
"grad_norm": 0.36757537722587585,
"learning_rate": 8.333333333333334e-05,
"loss": 0.0468,
"step": 65
},
{
"epoch": 0.28261742272179846,
"grad_norm": 0.33179935812950134,
"learning_rate": 8.284313725490198e-05,
"loss": 0.0601,
"step": 66
},
{
"epoch": 0.28689950488425,
"grad_norm": 0.40680983662605286,
"learning_rate": 8.23529411764706e-05,
"loss": 0.0565,
"step": 67
},
{
"epoch": 0.29118158704670144,
"grad_norm": 0.3238675594329834,
"learning_rate": 8.186274509803921e-05,
"loss": 0.0488,
"step": 68
},
{
"epoch": 0.29546366920915296,
"grad_norm": 0.39186710119247437,
"learning_rate": 8.137254901960785e-05,
"loss": 0.0505,
"step": 69
},
{
"epoch": 0.2997457513716044,
"grad_norm": 0.2939775288105011,
"learning_rate": 8.088235294117648e-05,
"loss": 0.0484,
"step": 70
},
{
"epoch": 0.30402783353405594,
"grad_norm": 0.37278881669044495,
"learning_rate": 8.039215686274511e-05,
"loss": 0.0504,
"step": 71
},
{
"epoch": 0.30830991569650745,
"grad_norm": 0.3936941921710968,
"learning_rate": 7.990196078431373e-05,
"loss": 0.055,
"step": 72
},
{
"epoch": 0.3125919978589589,
"grad_norm": 0.38660097122192383,
"learning_rate": 7.941176470588235e-05,
"loss": 0.0821,
"step": 73
},
{
"epoch": 0.31687408002141043,
"grad_norm": 0.30945825576782227,
"learning_rate": 7.892156862745098e-05,
"loss": 0.0442,
"step": 74
},
{
"epoch": 0.3211561621838619,
"grad_norm": 0.31412258744239807,
"learning_rate": 7.843137254901961e-05,
"loss": 0.0419,
"step": 75
},
{
"epoch": 0.3254382443463134,
"grad_norm": 0.34703952074050903,
"learning_rate": 7.794117647058824e-05,
"loss": 0.0575,
"step": 76
},
{
"epoch": 0.32972032650876487,
"grad_norm": 0.346296489238739,
"learning_rate": 7.745098039215687e-05,
"loss": 0.0603,
"step": 77
},
{
"epoch": 0.3340024086712164,
"grad_norm": 0.2502746284008026,
"learning_rate": 7.696078431372549e-05,
"loss": 0.0417,
"step": 78
},
{
"epoch": 0.33828449083366785,
"grad_norm": 0.3630567193031311,
"learning_rate": 7.647058823529411e-05,
"loss": 0.0591,
"step": 79
},
{
"epoch": 0.34256657299611937,
"grad_norm": 0.33694136142730713,
"learning_rate": 7.598039215686274e-05,
"loss": 0.047,
"step": 80
},
{
"epoch": 0.34684865515857083,
"grad_norm": 0.22548595070838928,
"learning_rate": 7.549019607843137e-05,
"loss": 0.0448,
"step": 81
},
{
"epoch": 0.35113073732102235,
"grad_norm": 0.3071158528327942,
"learning_rate": 7.500000000000001e-05,
"loss": 0.0503,
"step": 82
},
{
"epoch": 0.35541281948347386,
"grad_norm": 0.3624606132507324,
"learning_rate": 7.450980392156864e-05,
"loss": 0.0475,
"step": 83
},
{
"epoch": 0.3596949016459253,
"grad_norm": 0.23342928290367126,
"learning_rate": 7.401960784313726e-05,
"loss": 0.0337,
"step": 84
},
{
"epoch": 0.36397698380837684,
"grad_norm": 0.2770336866378784,
"learning_rate": 7.352941176470589e-05,
"loss": 0.0453,
"step": 85
},
{
"epoch": 0.3682590659708283,
"grad_norm": 0.27268415689468384,
"learning_rate": 7.303921568627451e-05,
"loss": 0.0508,
"step": 86
},
{
"epoch": 0.3725411481332798,
"grad_norm": 0.22464269399642944,
"learning_rate": 7.254901960784314e-05,
"loss": 0.0348,
"step": 87
},
{
"epoch": 0.3768232302957313,
"grad_norm": 0.27807021141052246,
"learning_rate": 7.205882352941177e-05,
"loss": 0.0597,
"step": 88
},
{
"epoch": 0.3811053124581828,
"grad_norm": 0.30591973662376404,
"learning_rate": 7.156862745098039e-05,
"loss": 0.0532,
"step": 89
},
{
"epoch": 0.38538739462063426,
"grad_norm": 0.29489806294441223,
"learning_rate": 7.107843137254902e-05,
"loss": 0.057,
"step": 90
},
{
"epoch": 0.3896694767830858,
"grad_norm": 0.27410319447517395,
"learning_rate": 7.058823529411765e-05,
"loss": 0.0529,
"step": 91
},
{
"epoch": 0.3939515589455373,
"grad_norm": 0.24769103527069092,
"learning_rate": 7.009803921568627e-05,
"loss": 0.0371,
"step": 92
},
{
"epoch": 0.39823364110798876,
"grad_norm": 0.30047643184661865,
"learning_rate": 6.96078431372549e-05,
"loss": 0.06,
"step": 93
},
{
"epoch": 0.4025157232704403,
"grad_norm": 0.2860408425331116,
"learning_rate": 6.911764705882354e-05,
"loss": 0.046,
"step": 94
},
{
"epoch": 0.40679780543289173,
"grad_norm": 0.2957577407360077,
"learning_rate": 6.862745098039216e-05,
"loss": 0.0585,
"step": 95
},
{
"epoch": 0.41107988759534325,
"grad_norm": 0.3263351023197174,
"learning_rate": 6.813725490196079e-05,
"loss": 0.0538,
"step": 96
},
{
"epoch": 0.4153619697577947,
"grad_norm": 0.35550421476364136,
"learning_rate": 6.764705882352942e-05,
"loss": 0.0533,
"step": 97
},
{
"epoch": 0.41964405192024623,
"grad_norm": 0.2950674295425415,
"learning_rate": 6.715686274509804e-05,
"loss": 0.05,
"step": 98
},
{
"epoch": 0.4239261340826977,
"grad_norm": 0.2513452172279358,
"learning_rate": 6.666666666666667e-05,
"loss": 0.0361,
"step": 99
},
{
"epoch": 0.4282082162451492,
"grad_norm": 0.3161787688732147,
"learning_rate": 6.61764705882353e-05,
"loss": 0.0483,
"step": 100
},
{
"epoch": 0.43249029840760067,
"grad_norm": 0.3478478789329529,
"learning_rate": 6.568627450980392e-05,
"loss": 0.0467,
"step": 101
},
{
"epoch": 0.4367723805700522,
"grad_norm": 0.33523473143577576,
"learning_rate": 6.519607843137255e-05,
"loss": 0.0593,
"step": 102
},
{
"epoch": 0.4410544627325037,
"grad_norm": 0.36270254850387573,
"learning_rate": 6.470588235294118e-05,
"loss": 0.061,
"step": 103
},
{
"epoch": 0.44533654489495517,
"grad_norm": 0.31555601954460144,
"learning_rate": 6.421568627450982e-05,
"loss": 0.0404,
"step": 104
},
{
"epoch": 0.4496186270574067,
"grad_norm": 0.2953006625175476,
"learning_rate": 6.372549019607843e-05,
"loss": 0.0502,
"step": 105
},
{
"epoch": 0.45390070921985815,
"grad_norm": 0.30984699726104736,
"learning_rate": 6.323529411764705e-05,
"loss": 0.0594,
"step": 106
},
{
"epoch": 0.45818279138230966,
"grad_norm": 0.30836424231529236,
"learning_rate": 6.274509803921569e-05,
"loss": 0.04,
"step": 107
},
{
"epoch": 0.4624648735447611,
"grad_norm": 0.27564123272895813,
"learning_rate": 6.225490196078432e-05,
"loss": 0.0372,
"step": 108
},
{
"epoch": 0.46674695570721264,
"grad_norm": 0.31601759791374207,
"learning_rate": 6.176470588235295e-05,
"loss": 0.0483,
"step": 109
},
{
"epoch": 0.4710290378696641,
"grad_norm": 0.40431514382362366,
"learning_rate": 6.127450980392158e-05,
"loss": 0.0531,
"step": 110
},
{
"epoch": 0.4753111200321156,
"grad_norm": 0.2956823408603668,
"learning_rate": 6.078431372549019e-05,
"loss": 0.0393,
"step": 111
},
{
"epoch": 0.4795932021945671,
"grad_norm": 0.26746317744255066,
"learning_rate": 6.0294117647058825e-05,
"loss": 0.0435,
"step": 112
},
{
"epoch": 0.4838752843570186,
"grad_norm": 0.29554516077041626,
"learning_rate": 5.980392156862745e-05,
"loss": 0.0419,
"step": 113
},
{
"epoch": 0.4881573665194701,
"grad_norm": 0.37722569704055786,
"learning_rate": 5.931372549019608e-05,
"loss": 0.0512,
"step": 114
},
{
"epoch": 0.4924394486819216,
"grad_norm": 0.2477983981370926,
"learning_rate": 5.882352941176471e-05,
"loss": 0.038,
"step": 115
},
{
"epoch": 0.4967215308443731,
"grad_norm": 0.22754943370819092,
"learning_rate": 5.833333333333334e-05,
"loss": 0.0377,
"step": 116
},
{
"epoch": 0.5010036130068246,
"grad_norm": 0.25772324204444885,
"learning_rate": 5.784313725490197e-05,
"loss": 0.036,
"step": 117
},
{
"epoch": 0.505285695169276,
"grad_norm": 0.2738616168498993,
"learning_rate": 5.735294117647059e-05,
"loss": 0.0581,
"step": 118
},
{
"epoch": 0.5095677773317275,
"grad_norm": 0.31121647357940674,
"learning_rate": 5.6862745098039215e-05,
"loss": 0.0479,
"step": 119
},
{
"epoch": 0.513849859494179,
"grad_norm": 0.24819445610046387,
"learning_rate": 5.637254901960785e-05,
"loss": 0.0375,
"step": 120
},
{
"epoch": 0.5181319416566306,
"grad_norm": 0.3565673232078552,
"learning_rate": 5.588235294117647e-05,
"loss": 0.0673,
"step": 121
},
{
"epoch": 0.522414023819082,
"grad_norm": 0.2924639880657196,
"learning_rate": 5.5392156862745104e-05,
"loss": 0.0439,
"step": 122
},
{
"epoch": 0.5266961059815335,
"grad_norm": 0.2511173486709595,
"learning_rate": 5.490196078431373e-05,
"loss": 0.043,
"step": 123
},
{
"epoch": 0.530978188143985,
"grad_norm": 0.3577595353126526,
"learning_rate": 5.441176470588235e-05,
"loss": 0.0573,
"step": 124
},
{
"epoch": 0.5352602703064365,
"grad_norm": 0.2690516412258148,
"learning_rate": 5.392156862745098e-05,
"loss": 0.0462,
"step": 125
},
{
"epoch": 0.539542352468888,
"grad_norm": 0.2612890303134918,
"learning_rate": 5.343137254901961e-05,
"loss": 0.0452,
"step": 126
},
{
"epoch": 0.5438244346313394,
"grad_norm": 0.27475517988204956,
"learning_rate": 5.294117647058824e-05,
"loss": 0.0436,
"step": 127
},
{
"epoch": 0.548106516793791,
"grad_norm": 0.33208557963371277,
"learning_rate": 5.245098039215687e-05,
"loss": 0.0508,
"step": 128
},
{
"epoch": 0.5523885989562425,
"grad_norm": 0.28704285621643066,
"learning_rate": 5.1960784313725495e-05,
"loss": 0.0436,
"step": 129
},
{
"epoch": 0.556670681118694,
"grad_norm": 0.35237687826156616,
"learning_rate": 5.147058823529411e-05,
"loss": 0.0549,
"step": 130
},
{
"epoch": 0.5609527632811454,
"grad_norm": 0.290622740983963,
"learning_rate": 5.0980392156862745e-05,
"loss": 0.048,
"step": 131
},
{
"epoch": 0.5652348454435969,
"grad_norm": 0.250241219997406,
"learning_rate": 5.049019607843137e-05,
"loss": 0.0441,
"step": 132
},
{
"epoch": 0.5695169276060484,
"grad_norm": 0.3002930283546448,
"learning_rate": 5e-05,
"loss": 0.0411,
"step": 133
},
{
"epoch": 0.5737990097685,
"grad_norm": 0.30466729402542114,
"learning_rate": 4.9509803921568634e-05,
"loss": 0.0452,
"step": 134
},
{
"epoch": 0.5780810919309515,
"grad_norm": 0.2513613998889923,
"learning_rate": 4.901960784313725e-05,
"loss": 0.0408,
"step": 135
},
{
"epoch": 0.5823631740934029,
"grad_norm": 0.30394110083580017,
"learning_rate": 4.8529411764705885e-05,
"loss": 0.0451,
"step": 136
},
{
"epoch": 0.5866452562558544,
"grad_norm": 0.2374386042356491,
"learning_rate": 4.803921568627452e-05,
"loss": 0.0391,
"step": 137
},
{
"epoch": 0.5909273384183059,
"grad_norm": 0.3214227259159088,
"learning_rate": 4.7549019607843135e-05,
"loss": 0.0428,
"step": 138
},
{
"epoch": 0.5952094205807574,
"grad_norm": 0.25820499658584595,
"learning_rate": 4.705882352941177e-05,
"loss": 0.0458,
"step": 139
},
{
"epoch": 0.5994915027432088,
"grad_norm": 0.21727867424488068,
"learning_rate": 4.656862745098039e-05,
"loss": 0.0313,
"step": 140
},
{
"epoch": 0.6037735849056604,
"grad_norm": 0.3863449692726135,
"learning_rate": 4.607843137254902e-05,
"loss": 0.0578,
"step": 141
},
{
"epoch": 0.6080556670681119,
"grad_norm": 0.2634093463420868,
"learning_rate": 4.558823529411765e-05,
"loss": 0.0373,
"step": 142
},
{
"epoch": 0.6123377492305634,
"grad_norm": 0.2833387553691864,
"learning_rate": 4.5098039215686275e-05,
"loss": 0.051,
"step": 143
},
{
"epoch": 0.6166198313930149,
"grad_norm": 0.3331435024738312,
"learning_rate": 4.460784313725491e-05,
"loss": 0.0639,
"step": 144
},
{
"epoch": 0.6209019135554663,
"grad_norm": 0.2673349380493164,
"learning_rate": 4.411764705882353e-05,
"loss": 0.0509,
"step": 145
},
{
"epoch": 0.6251839957179178,
"grad_norm": 0.3010377585887909,
"learning_rate": 4.362745098039216e-05,
"loss": 0.0521,
"step": 146
},
{
"epoch": 0.6294660778803693,
"grad_norm": 0.28252658247947693,
"learning_rate": 4.313725490196079e-05,
"loss": 0.0455,
"step": 147
},
{
"epoch": 0.6337481600428209,
"grad_norm": 0.28520438075065613,
"learning_rate": 4.2647058823529415e-05,
"loss": 0.0551,
"step": 148
},
{
"epoch": 0.6380302422052723,
"grad_norm": 0.24448159337043762,
"learning_rate": 4.215686274509804e-05,
"loss": 0.0362,
"step": 149
},
{
"epoch": 0.6423123243677238,
"grad_norm": 0.2240142524242401,
"learning_rate": 4.166666666666667e-05,
"loss": 0.0377,
"step": 150
},
{
"epoch": 0.6465944065301753,
"grad_norm": 0.2896118760108948,
"learning_rate": 4.11764705882353e-05,
"loss": 0.0378,
"step": 151
},
{
"epoch": 0.6508764886926268,
"grad_norm": 0.2663087844848633,
"learning_rate": 4.068627450980392e-05,
"loss": 0.0407,
"step": 152
},
{
"epoch": 0.6551585708550783,
"grad_norm": 0.3086792230606079,
"learning_rate": 4.0196078431372555e-05,
"loss": 0.0323,
"step": 153
},
{
"epoch": 0.6594406530175297,
"grad_norm": 0.2776995897293091,
"learning_rate": 3.970588235294117e-05,
"loss": 0.0437,
"step": 154
},
{
"epoch": 0.6637227351799813,
"grad_norm": 0.3368375897407532,
"learning_rate": 3.9215686274509805e-05,
"loss": 0.0525,
"step": 155
},
{
"epoch": 0.6680048173424328,
"grad_norm": 0.25399115681648254,
"learning_rate": 3.872549019607844e-05,
"loss": 0.0352,
"step": 156
},
{
"epoch": 0.6722868995048843,
"grad_norm": 0.3201920986175537,
"learning_rate": 3.8235294117647055e-05,
"loss": 0.046,
"step": 157
},
{
"epoch": 0.6765689816673357,
"grad_norm": 0.36324596405029297,
"learning_rate": 3.774509803921569e-05,
"loss": 0.0429,
"step": 158
},
{
"epoch": 0.6808510638297872,
"grad_norm": 0.27535951137542725,
"learning_rate": 3.725490196078432e-05,
"loss": 0.036,
"step": 159
},
{
"epoch": 0.6851331459922387,
"grad_norm": 0.2501991391181946,
"learning_rate": 3.6764705882352945e-05,
"loss": 0.0333,
"step": 160
},
{
"epoch": 0.6894152281546903,
"grad_norm": 0.26270344853401184,
"learning_rate": 3.627450980392157e-05,
"loss": 0.0359,
"step": 161
},
{
"epoch": 0.6936973103171417,
"grad_norm": 0.2966891825199127,
"learning_rate": 3.5784313725490195e-05,
"loss": 0.0448,
"step": 162
},
{
"epoch": 0.6979793924795932,
"grad_norm": 0.24465243518352509,
"learning_rate": 3.529411764705883e-05,
"loss": 0.0488,
"step": 163
},
{
"epoch": 0.7022614746420447,
"grad_norm": 0.2697572410106659,
"learning_rate": 3.480392156862745e-05,
"loss": 0.0295,
"step": 164
},
{
"epoch": 0.7065435568044962,
"grad_norm": 0.27013951539993286,
"learning_rate": 3.431372549019608e-05,
"loss": 0.0395,
"step": 165
},
{
"epoch": 0.7108256389669477,
"grad_norm": 0.3045112192630768,
"learning_rate": 3.382352941176471e-05,
"loss": 0.0411,
"step": 166
},
{
"epoch": 0.7151077211293991,
"grad_norm": 0.24719510972499847,
"learning_rate": 3.3333333333333335e-05,
"loss": 0.0354,
"step": 167
},
{
"epoch": 0.7193898032918506,
"grad_norm": 0.1892707198858261,
"learning_rate": 3.284313725490196e-05,
"loss": 0.0234,
"step": 168
},
{
"epoch": 0.7236718854543022,
"grad_norm": 0.30045631527900696,
"learning_rate": 3.235294117647059e-05,
"loss": 0.0447,
"step": 169
},
{
"epoch": 0.7279539676167537,
"grad_norm": 0.309243768453598,
"learning_rate": 3.186274509803922e-05,
"loss": 0.0404,
"step": 170
},
{
"epoch": 0.7322360497792051,
"grad_norm": 0.24024784564971924,
"learning_rate": 3.137254901960784e-05,
"loss": 0.0395,
"step": 171
},
{
"epoch": 0.7365181319416566,
"grad_norm": 0.2552880048751831,
"learning_rate": 3.0882352941176475e-05,
"loss": 0.0359,
"step": 172
},
{
"epoch": 0.7408002141041081,
"grad_norm": 0.33713486790657043,
"learning_rate": 3.0392156862745097e-05,
"loss": 0.0614,
"step": 173
},
{
"epoch": 0.7450822962665596,
"grad_norm": 0.2314293533563614,
"learning_rate": 2.9901960784313725e-05,
"loss": 0.0348,
"step": 174
},
{
"epoch": 0.7493643784290112,
"grad_norm": 0.23023654520511627,
"learning_rate": 2.9411764705882354e-05,
"loss": 0.0419,
"step": 175
},
{
"epoch": 0.7536464605914626,
"grad_norm": 0.2514519691467285,
"learning_rate": 2.8921568627450986e-05,
"loss": 0.0388,
"step": 176
},
{
"epoch": 0.7579285427539141,
"grad_norm": 0.2647082209587097,
"learning_rate": 2.8431372549019608e-05,
"loss": 0.0463,
"step": 177
},
{
"epoch": 0.7622106249163656,
"grad_norm": 0.29537469148635864,
"learning_rate": 2.7941176470588236e-05,
"loss": 0.0352,
"step": 178
},
{
"epoch": 0.7664927070788171,
"grad_norm": 0.2043161392211914,
"learning_rate": 2.7450980392156865e-05,
"loss": 0.0387,
"step": 179
},
{
"epoch": 0.7707747892412685,
"grad_norm": 0.305243581533432,
"learning_rate": 2.696078431372549e-05,
"loss": 0.0346,
"step": 180
},
{
"epoch": 0.77505687140372,
"grad_norm": 0.29974740743637085,
"learning_rate": 2.647058823529412e-05,
"loss": 0.0513,
"step": 181
},
{
"epoch": 0.7793389535661716,
"grad_norm": 0.28913021087646484,
"learning_rate": 2.5980392156862747e-05,
"loss": 0.0341,
"step": 182
},
{
"epoch": 0.7836210357286231,
"grad_norm": 0.32380354404449463,
"learning_rate": 2.5490196078431373e-05,
"loss": 0.0614,
"step": 183
},
{
"epoch": 0.7879031178910746,
"grad_norm": 0.2799640893936157,
"learning_rate": 2.5e-05,
"loss": 0.0515,
"step": 184
},
{
"epoch": 0.792185200053526,
"grad_norm": 0.28914907574653625,
"learning_rate": 2.4509803921568626e-05,
"loss": 0.0393,
"step": 185
},
{
"epoch": 0.7964672822159775,
"grad_norm": 0.26201391220092773,
"learning_rate": 2.401960784313726e-05,
"loss": 0.0431,
"step": 186
},
{
"epoch": 0.800749364378429,
"grad_norm": 0.29834750294685364,
"learning_rate": 2.3529411764705884e-05,
"loss": 0.039,
"step": 187
},
{
"epoch": 0.8050314465408805,
"grad_norm": 0.28359827399253845,
"learning_rate": 2.303921568627451e-05,
"loss": 0.0474,
"step": 188
},
{
"epoch": 0.809313528703332,
"grad_norm": 0.21148712933063507,
"learning_rate": 2.2549019607843138e-05,
"loss": 0.0355,
"step": 189
},
{
"epoch": 0.8135956108657835,
"grad_norm": 0.30203911662101746,
"learning_rate": 2.2058823529411766e-05,
"loss": 0.0423,
"step": 190
},
{
"epoch": 0.817877693028235,
"grad_norm": 0.32118844985961914,
"learning_rate": 2.1568627450980395e-05,
"loss": 0.0396,
"step": 191
},
{
"epoch": 0.8221597751906865,
"grad_norm": 0.2358664572238922,
"learning_rate": 2.107843137254902e-05,
"loss": 0.0344,
"step": 192
},
{
"epoch": 0.8264418573531379,
"grad_norm": 0.3052966892719269,
"learning_rate": 2.058823529411765e-05,
"loss": 0.0459,
"step": 193
},
{
"epoch": 0.8307239395155894,
"grad_norm": 0.2716437578201294,
"learning_rate": 2.0098039215686277e-05,
"loss": 0.0511,
"step": 194
},
{
"epoch": 0.8350060216780409,
"grad_norm": 0.3086344301700592,
"learning_rate": 1.9607843137254903e-05,
"loss": 0.0477,
"step": 195
},
{
"epoch": 0.8392881038404925,
"grad_norm": 0.26460692286491394,
"learning_rate": 1.9117647058823528e-05,
"loss": 0.0355,
"step": 196
},
{
"epoch": 0.843570186002944,
"grad_norm": 0.2656954526901245,
"learning_rate": 1.862745098039216e-05,
"loss": 0.0504,
"step": 197
},
{
"epoch": 0.8478522681653954,
"grad_norm": 0.2059815526008606,
"learning_rate": 1.8137254901960785e-05,
"loss": 0.0404,
"step": 198
},
{
"epoch": 0.8521343503278469,
"grad_norm": 0.28695958852767944,
"learning_rate": 1.7647058823529414e-05,
"loss": 0.0502,
"step": 199
},
{
"epoch": 0.8564164324902984,
"grad_norm": 0.278156042098999,
"learning_rate": 1.715686274509804e-05,
"loss": 0.0325,
"step": 200
},
{
"epoch": 0.8606985146527499,
"grad_norm": 0.33728834986686707,
"learning_rate": 1.6666666666666667e-05,
"loss": 0.0468,
"step": 201
},
{
"epoch": 0.8649805968152013,
"grad_norm": 0.2665461003780365,
"learning_rate": 1.6176470588235296e-05,
"loss": 0.041,
"step": 202
},
{
"epoch": 0.8692626789776529,
"grad_norm": 0.2768334150314331,
"learning_rate": 1.568627450980392e-05,
"loss": 0.0477,
"step": 203
},
{
"epoch": 0.8735447611401044,
"grad_norm": 0.2847307324409485,
"learning_rate": 1.5196078431372548e-05,
"loss": 0.0368,
"step": 204
},
{
"epoch": 0.8778268433025559,
"grad_norm": 0.26865965127944946,
"learning_rate": 1.4705882352941177e-05,
"loss": 0.0321,
"step": 205
},
{
"epoch": 0.8821089254650074,
"grad_norm": 0.23639728128910065,
"learning_rate": 1.4215686274509804e-05,
"loss": 0.0267,
"step": 206
},
{
"epoch": 0.8863910076274588,
"grad_norm": 0.22090226411819458,
"learning_rate": 1.3725490196078432e-05,
"loss": 0.0283,
"step": 207
},
{
"epoch": 0.8906730897899103,
"grad_norm": 0.24385924637317657,
"learning_rate": 1.323529411764706e-05,
"loss": 0.0486,
"step": 208
},
{
"epoch": 0.8949551719523618,
"grad_norm": 0.267991304397583,
"learning_rate": 1.2745098039215686e-05,
"loss": 0.049,
"step": 209
},
{
"epoch": 0.8992372541148134,
"grad_norm": 0.2858865261077881,
"learning_rate": 1.2254901960784313e-05,
"loss": 0.0416,
"step": 210
},
{
"epoch": 0.9035193362772648,
"grad_norm": 0.30461645126342773,
"learning_rate": 1.1764705882352942e-05,
"loss": 0.0459,
"step": 211
},
{
"epoch": 0.9078014184397163,
"grad_norm": 0.3148707449436188,
"learning_rate": 1.1274509803921569e-05,
"loss": 0.0448,
"step": 212
},
{
"epoch": 0.9120835006021678,
"grad_norm": 0.29912617802619934,
"learning_rate": 1.0784313725490197e-05,
"loss": 0.0391,
"step": 213
},
{
"epoch": 0.9163655827646193,
"grad_norm": 0.30762436985969543,
"learning_rate": 1.0294117647058824e-05,
"loss": 0.0475,
"step": 214
},
{
"epoch": 0.9206476649270708,
"grad_norm": 0.27570030093193054,
"learning_rate": 9.803921568627451e-06,
"loss": 0.0371,
"step": 215
},
{
"epoch": 0.9249297470895222,
"grad_norm": 0.26668745279312134,
"learning_rate": 9.31372549019608e-06,
"loss": 0.0532,
"step": 216
},
{
"epoch": 0.9292118292519738,
"grad_norm": 0.20003274083137512,
"learning_rate": 8.823529411764707e-06,
"loss": 0.0328,
"step": 217
},
{
"epoch": 0.9334939114144253,
"grad_norm": 0.23162154853343964,
"learning_rate": 8.333333333333334e-06,
"loss": 0.0348,
"step": 218
},
{
"epoch": 0.9377759935768768,
"grad_norm": 0.28092819452285767,
"learning_rate": 7.84313725490196e-06,
"loss": 0.041,
"step": 219
},
{
"epoch": 0.9420580757393282,
"grad_norm": 0.20697717368602753,
"learning_rate": 7.3529411764705884e-06,
"loss": 0.0344,
"step": 220
},
{
"epoch": 0.9463401579017797,
"grad_norm": 0.2537749111652374,
"learning_rate": 6.862745098039216e-06,
"loss": 0.0364,
"step": 221
},
{
"epoch": 0.9506222400642312,
"grad_norm": 0.23425333201885223,
"learning_rate": 6.372549019607843e-06,
"loss": 0.0391,
"step": 222
},
{
"epoch": 0.9549043222266828,
"grad_norm": 0.25059816241264343,
"learning_rate": 5.882352941176471e-06,
"loss": 0.0506,
"step": 223
},
{
"epoch": 0.9591864043891342,
"grad_norm": 0.2436205893754959,
"learning_rate": 5.392156862745099e-06,
"loss": 0.0418,
"step": 224
},
{
"epoch": 0.9634684865515857,
"grad_norm": 0.255195289850235,
"learning_rate": 4.901960784313726e-06,
"loss": 0.0412,
"step": 225
},
{
"epoch": 0.9677505687140372,
"grad_norm": 0.27884790301322937,
"learning_rate": 4.411764705882353e-06,
"loss": 0.0497,
"step": 226
},
{
"epoch": 0.9720326508764887,
"grad_norm": 0.25363633036613464,
"learning_rate": 3.92156862745098e-06,
"loss": 0.0408,
"step": 227
},
{
"epoch": 0.9763147330389402,
"grad_norm": 0.30385518074035645,
"learning_rate": 3.431372549019608e-06,
"loss": 0.0552,
"step": 228
},
{
"epoch": 0.9805968152013916,
"grad_norm": 0.27926793694496155,
"learning_rate": 2.9411764705882355e-06,
"loss": 0.036,
"step": 229
},
{
"epoch": 0.9848788973638432,
"grad_norm": 0.236245259642601,
"learning_rate": 2.450980392156863e-06,
"loss": 0.029,
"step": 230
},
{
"epoch": 0.9891609795262947,
"grad_norm": 0.25143948197364807,
"learning_rate": 1.96078431372549e-06,
"loss": 0.0303,
"step": 231
},
{
"epoch": 0.9934430616887462,
"grad_norm": 0.29792022705078125,
"learning_rate": 1.4705882352941177e-06,
"loss": 0.0498,
"step": 232
},
{
"epoch": 0.9977251438511976,
"grad_norm": 0.32305002212524414,
"learning_rate": 9.80392156862745e-07,
"loss": 0.0375,
"step": 233
},
{
"epoch": 1.0,
"grad_norm": 0.4602526128292084,
"learning_rate": 4.901960784313725e-07,
"loss": 0.0426,
"step": 234
}
],
"logging_steps": 1,
"max_steps": 234,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 5000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.927681385546916e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}