Files
openthoughts3_100k_qwen25_1…/trainer_state.json
2025-10-01 02:18:26 +00:00

3438 lines
83 KiB
JSON

{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.961636828644501,
"eval_steps": 500,
"global_step": 485,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.010230179028132993,
"grad_norm": 2.9682868161196363,
"learning_rate": 3.2653061224489794e-06,
"loss": 1.4918,
"step": 1
},
{
"epoch": 0.020460358056265986,
"grad_norm": 2.9796653677109557,
"learning_rate": 6.530612244897959e-06,
"loss": 1.4922,
"step": 2
},
{
"epoch": 0.030690537084398978,
"grad_norm": 2.8660911829024167,
"learning_rate": 9.795918367346939e-06,
"loss": 1.503,
"step": 3
},
{
"epoch": 0.04092071611253197,
"grad_norm": 1.9805780511052813,
"learning_rate": 1.3061224489795918e-05,
"loss": 1.4675,
"step": 4
},
{
"epoch": 0.05115089514066496,
"grad_norm": 1.998834758427541,
"learning_rate": 1.63265306122449e-05,
"loss": 1.4533,
"step": 5
},
{
"epoch": 0.061381074168797956,
"grad_norm": 2.050049149995988,
"learning_rate": 1.9591836734693877e-05,
"loss": 1.4122,
"step": 6
},
{
"epoch": 0.07161125319693094,
"grad_norm": 2.399845329221154,
"learning_rate": 2.2857142857142858e-05,
"loss": 1.4056,
"step": 7
},
{
"epoch": 0.08184143222506395,
"grad_norm": 1.7232075850783724,
"learning_rate": 2.6122448979591835e-05,
"loss": 1.3833,
"step": 8
},
{
"epoch": 0.09207161125319693,
"grad_norm": 1.410904151553637,
"learning_rate": 2.938775510204082e-05,
"loss": 1.3843,
"step": 9
},
{
"epoch": 0.10230179028132992,
"grad_norm": 1.1900105733063466,
"learning_rate": 3.26530612244898e-05,
"loss": 1.3515,
"step": 10
},
{
"epoch": 0.11253196930946291,
"grad_norm": 1.2160713923639859,
"learning_rate": 3.591836734693878e-05,
"loss": 1.3224,
"step": 11
},
{
"epoch": 0.12276214833759591,
"grad_norm": 1.0389820905559795,
"learning_rate": 3.9183673469387755e-05,
"loss": 1.322,
"step": 12
},
{
"epoch": 0.1329923273657289,
"grad_norm": 0.9606384846443378,
"learning_rate": 4.244897959183674e-05,
"loss": 1.2956,
"step": 13
},
{
"epoch": 0.1432225063938619,
"grad_norm": 1.032658088356086,
"learning_rate": 4.5714285714285716e-05,
"loss": 1.3067,
"step": 14
},
{
"epoch": 0.1534526854219949,
"grad_norm": 1.006086286942022,
"learning_rate": 4.89795918367347e-05,
"loss": 1.2961,
"step": 15
},
{
"epoch": 0.1636828644501279,
"grad_norm": 1.1671471070443895,
"learning_rate": 5.224489795918367e-05,
"loss": 1.301,
"step": 16
},
{
"epoch": 0.17391304347826086,
"grad_norm": 0.8672174854035504,
"learning_rate": 5.551020408163266e-05,
"loss": 1.2656,
"step": 17
},
{
"epoch": 0.18414322250639387,
"grad_norm": 0.7154082384610629,
"learning_rate": 5.877551020408164e-05,
"loss": 1.246,
"step": 18
},
{
"epoch": 0.19437340153452684,
"grad_norm": 0.694176256293125,
"learning_rate": 6.204081632653062e-05,
"loss": 1.2565,
"step": 19
},
{
"epoch": 0.20460358056265984,
"grad_norm": 0.7538553851540746,
"learning_rate": 6.53061224489796e-05,
"loss": 1.241,
"step": 20
},
{
"epoch": 0.21483375959079284,
"grad_norm": 0.7434156335821637,
"learning_rate": 6.857142857142857e-05,
"loss": 1.232,
"step": 21
},
{
"epoch": 0.22506393861892582,
"grad_norm": 1.2022520877750795,
"learning_rate": 7.183673469387756e-05,
"loss": 1.2567,
"step": 22
},
{
"epoch": 0.23529411764705882,
"grad_norm": 1.150528684217217,
"learning_rate": 7.510204081632654e-05,
"loss": 1.2448,
"step": 23
},
{
"epoch": 0.24552429667519182,
"grad_norm": 0.8365697619839101,
"learning_rate": 7.836734693877551e-05,
"loss": 1.2357,
"step": 24
},
{
"epoch": 0.2557544757033248,
"grad_norm": 1.1469314213901645,
"learning_rate": 8.16326530612245e-05,
"loss": 1.2164,
"step": 25
},
{
"epoch": 0.2659846547314578,
"grad_norm": 0.9677681397432715,
"learning_rate": 8.489795918367348e-05,
"loss": 1.2115,
"step": 26
},
{
"epoch": 0.27621483375959077,
"grad_norm": 1.106826210998024,
"learning_rate": 8.816326530612245e-05,
"loss": 1.2004,
"step": 27
},
{
"epoch": 0.2864450127877238,
"grad_norm": 1.1744803496009477,
"learning_rate": 9.142857142857143e-05,
"loss": 1.2235,
"step": 28
},
{
"epoch": 0.2966751918158568,
"grad_norm": 1.3610877277166293,
"learning_rate": 9.469387755102041e-05,
"loss": 1.208,
"step": 29
},
{
"epoch": 0.3069053708439898,
"grad_norm": 0.7447757731907216,
"learning_rate": 9.79591836734694e-05,
"loss": 1.2118,
"step": 30
},
{
"epoch": 0.3171355498721228,
"grad_norm": 1.3296162158652725,
"learning_rate": 0.00010122448979591839,
"loss": 1.2032,
"step": 31
},
{
"epoch": 0.3273657289002558,
"grad_norm": 1.144264777414957,
"learning_rate": 0.00010448979591836734,
"loss": 1.1912,
"step": 32
},
{
"epoch": 0.3375959079283887,
"grad_norm": 1.0342650793984924,
"learning_rate": 0.00010775510204081634,
"loss": 1.2025,
"step": 33
},
{
"epoch": 0.34782608695652173,
"grad_norm": 1.9336445124108894,
"learning_rate": 0.00011102040816326532,
"loss": 1.2118,
"step": 34
},
{
"epoch": 0.35805626598465473,
"grad_norm": 0.9195357493046576,
"learning_rate": 0.0001142857142857143,
"loss": 1.1986,
"step": 35
},
{
"epoch": 0.36828644501278773,
"grad_norm": 1.588300333871633,
"learning_rate": 0.00011755102040816328,
"loss": 1.1939,
"step": 36
},
{
"epoch": 0.37851662404092073,
"grad_norm": 1.4710948512542918,
"learning_rate": 0.00012081632653061224,
"loss": 1.1997,
"step": 37
},
{
"epoch": 0.3887468030690537,
"grad_norm": 1.2164757524877552,
"learning_rate": 0.00012408163265306124,
"loss": 1.1791,
"step": 38
},
{
"epoch": 0.3989769820971867,
"grad_norm": 1.4375602427525511,
"learning_rate": 0.0001273469387755102,
"loss": 1.1896,
"step": 39
},
{
"epoch": 0.4092071611253197,
"grad_norm": 1.5736548612793597,
"learning_rate": 0.0001306122448979592,
"loss": 1.1876,
"step": 40
},
{
"epoch": 0.4194373401534527,
"grad_norm": 1.0735537900578096,
"learning_rate": 0.00013387755102040817,
"loss": 1.1709,
"step": 41
},
{
"epoch": 0.4296675191815857,
"grad_norm": 1.8510870420688348,
"learning_rate": 0.00013714285714285713,
"loss": 1.1795,
"step": 42
},
{
"epoch": 0.4398976982097187,
"grad_norm": 1.1080619334489261,
"learning_rate": 0.00014040816326530613,
"loss": 1.1753,
"step": 43
},
{
"epoch": 0.45012787723785164,
"grad_norm": 1.9521299298489614,
"learning_rate": 0.00014367346938775512,
"loss": 1.2028,
"step": 44
},
{
"epoch": 0.46035805626598464,
"grad_norm": 1.41657060512908,
"learning_rate": 0.0001469387755102041,
"loss": 1.1822,
"step": 45
},
{
"epoch": 0.47058823529411764,
"grad_norm": 1.1977389520948274,
"learning_rate": 0.00015020408163265308,
"loss": 1.1836,
"step": 46
},
{
"epoch": 0.48081841432225064,
"grad_norm": 1.5288668435257222,
"learning_rate": 0.00015346938775510205,
"loss": 1.1901,
"step": 47
},
{
"epoch": 0.49104859335038364,
"grad_norm": 0.9552067161901755,
"learning_rate": 0.00015673469387755102,
"loss": 1.1562,
"step": 48
},
{
"epoch": 0.5012787723785166,
"grad_norm": 1.6547871414680237,
"learning_rate": 0.00016,
"loss": 1.1794,
"step": 49
},
{
"epoch": 0.5115089514066496,
"grad_norm": 1.8368364441109388,
"learning_rate": 0.00015999792324684382,
"loss": 1.1877,
"step": 50
},
{
"epoch": 0.5217391304347826,
"grad_norm": 1.1718896076543888,
"learning_rate": 0.00015999169309519789,
"loss": 1.1585,
"step": 51
},
{
"epoch": 0.5319693094629157,
"grad_norm": 2.015243385621693,
"learning_rate": 0.0001599813098685243,
"loss": 1.1732,
"step": 52
},
{
"epoch": 0.5421994884910486,
"grad_norm": 1.638588399609178,
"learning_rate": 0.0001599667741059081,
"loss": 1.1617,
"step": 53
},
{
"epoch": 0.5524296675191815,
"grad_norm": 1.1352479883236335,
"learning_rate": 0.00015994808656202904,
"loss": 1.1584,
"step": 54
},
{
"epoch": 0.5626598465473146,
"grad_norm": 1.2074962340041349,
"learning_rate": 0.00015992524820712252,
"loss": 1.1478,
"step": 55
},
{
"epoch": 0.5728900255754475,
"grad_norm": 1.4410034215063858,
"learning_rate": 0.00015989826022692918,
"loss": 1.1493,
"step": 56
},
{
"epoch": 0.5831202046035806,
"grad_norm": 0.8679551934822172,
"learning_rate": 0.00015986712402263334,
"loss": 1.148,
"step": 57
},
{
"epoch": 0.5933503836317136,
"grad_norm": 1.3034239933412972,
"learning_rate": 0.00015983184121079024,
"loss": 1.1532,
"step": 58
},
{
"epoch": 0.6035805626598465,
"grad_norm": 0.9780209051309727,
"learning_rate": 0.00015979241362324223,
"loss": 1.1313,
"step": 59
},
{
"epoch": 0.6138107416879796,
"grad_norm": 1.057690857573897,
"learning_rate": 0.0001597488433070234,
"loss": 1.138,
"step": 60
},
{
"epoch": 0.6240409207161125,
"grad_norm": 1.456389761257225,
"learning_rate": 0.00015970113252425356,
"loss": 1.1546,
"step": 61
},
{
"epoch": 0.6342710997442456,
"grad_norm": 2.3692148862526485,
"learning_rate": 0.00015964928375202068,
"loss": 1.163,
"step": 62
},
{
"epoch": 0.6445012787723785,
"grad_norm": 1.047447123295706,
"learning_rate": 0.00015959329968225232,
"loss": 1.1564,
"step": 63
},
{
"epoch": 0.6547314578005116,
"grad_norm": 3.5383718747156534,
"learning_rate": 0.0001595331832215758,
"loss": 1.1684,
"step": 64
},
{
"epoch": 0.6649616368286445,
"grad_norm": 2.968800505278846,
"learning_rate": 0.00015946893749116734,
"loss": 1.1788,
"step": 65
},
{
"epoch": 0.6751918158567775,
"grad_norm": 1.9185023523212743,
"learning_rate": 0.00015940056582659006,
"loss": 1.1537,
"step": 66
},
{
"epoch": 0.6854219948849105,
"grad_norm": 1.404723257524876,
"learning_rate": 0.0001593280717776207,
"loss": 1.1487,
"step": 67
},
{
"epoch": 0.6956521739130435,
"grad_norm": 1.9443271708932357,
"learning_rate": 0.0001592514591080654,
"loss": 1.15,
"step": 68
},
{
"epoch": 0.7058823529411765,
"grad_norm": 1.0325757666109745,
"learning_rate": 0.0001591707317955642,
"loss": 1.1487,
"step": 69
},
{
"epoch": 0.7161125319693095,
"grad_norm": 2.1150324509757006,
"learning_rate": 0.00015908589403138468,
"loss": 1.1766,
"step": 70
},
{
"epoch": 0.7263427109974424,
"grad_norm": 1.7663965705197828,
"learning_rate": 0.00015899695022020415,
"loss": 1.1464,
"step": 71
},
{
"epoch": 0.7365728900255755,
"grad_norm": 1.4582930405137742,
"learning_rate": 0.00015890390497988116,
"loss": 1.1487,
"step": 72
},
{
"epoch": 0.7468030690537084,
"grad_norm": 1.394985742880279,
"learning_rate": 0.0001588067631412156,
"loss": 1.1447,
"step": 73
},
{
"epoch": 0.7570332480818415,
"grad_norm": 1.1897882316318342,
"learning_rate": 0.000158705529747698,
"loss": 1.1299,
"step": 74
},
{
"epoch": 0.7672634271099744,
"grad_norm": 1.1626082561212874,
"learning_rate": 0.0001586002100552476,
"loss": 1.1473,
"step": 75
},
{
"epoch": 0.7774936061381074,
"grad_norm": 1.0736945268719569,
"learning_rate": 0.00015849080953193943,
"loss": 1.1368,
"step": 76
},
{
"epoch": 0.7877237851662404,
"grad_norm": 1.5654343283068946,
"learning_rate": 0.00015837733385772062,
"loss": 1.123,
"step": 77
},
{
"epoch": 0.7979539641943734,
"grad_norm": 0.8523221287975669,
"learning_rate": 0.00015825978892411522,
"loss": 1.1403,
"step": 78
},
{
"epoch": 0.8081841432225064,
"grad_norm": 1.035617914076512,
"learning_rate": 0.00015813818083391858,
"loss": 1.1229,
"step": 79
},
{
"epoch": 0.8184143222506394,
"grad_norm": 1.4757207346915424,
"learning_rate": 0.0001580125159008803,
"loss": 1.129,
"step": 80
},
{
"epoch": 0.8286445012787724,
"grad_norm": 1.0711086873355355,
"learning_rate": 0.00015788280064937655,
"loss": 1.1351,
"step": 81
},
{
"epoch": 0.8388746803069054,
"grad_norm": 1.2110898160193528,
"learning_rate": 0.00015774904181407127,
"loss": 1.1334,
"step": 82
},
{
"epoch": 0.8491048593350383,
"grad_norm": 1.2411189602993498,
"learning_rate": 0.00015761124633956652,
"loss": 1.1363,
"step": 83
},
{
"epoch": 0.8593350383631714,
"grad_norm": 1.1341606871236802,
"learning_rate": 0.00015746942138004203,
"loss": 1.1142,
"step": 84
},
{
"epoch": 0.8695652173913043,
"grad_norm": 0.9942900226210764,
"learning_rate": 0.00015732357429888355,
"loss": 1.13,
"step": 85
},
{
"epoch": 0.8797953964194374,
"grad_norm": 1.1602205285909497,
"learning_rate": 0.00015717371266830076,
"loss": 1.1225,
"step": 86
},
{
"epoch": 0.8900255754475703,
"grad_norm": 1.2773792554197163,
"learning_rate": 0.000157019844268934,
"loss": 1.1237,
"step": 87
},
{
"epoch": 0.9002557544757033,
"grad_norm": 1.0698753174294156,
"learning_rate": 0.0001568619770894504,
"loss": 1.1223,
"step": 88
},
{
"epoch": 0.9104859335038363,
"grad_norm": 1.1279028063380314,
"learning_rate": 0.000156700119326129,
"loss": 1.1117,
"step": 89
},
{
"epoch": 0.9207161125319693,
"grad_norm": 1.047346608848377,
"learning_rate": 0.00015653427938243532,
"loss": 1.1195,
"step": 90
},
{
"epoch": 0.9309462915601023,
"grad_norm": 1.1063879184523018,
"learning_rate": 0.0001563644658685851,
"loss": 1.1112,
"step": 91
},
{
"epoch": 0.9411764705882353,
"grad_norm": 1.403473791127022,
"learning_rate": 0.00015619068760109703,
"loss": 1.1334,
"step": 92
},
{
"epoch": 0.9514066496163683,
"grad_norm": 1.1738079096105647,
"learning_rate": 0.00015601295360233528,
"loss": 1.123,
"step": 93
},
{
"epoch": 0.9616368286445013,
"grad_norm": 1.2622400384084083,
"learning_rate": 0.0001558312731000409,
"loss": 1.1245,
"step": 94
},
{
"epoch": 0.9718670076726342,
"grad_norm": 1.5103980615396817,
"learning_rate": 0.00015564565552685282,
"loss": 1.1159,
"step": 95
},
{
"epoch": 0.9820971867007673,
"grad_norm": 0.9924832092739461,
"learning_rate": 0.00015545611051981807,
"loss": 1.1086,
"step": 96
},
{
"epoch": 0.9923273657289002,
"grad_norm": 1.2306800508291864,
"learning_rate": 0.00015526264791989144,
"loss": 1.1396,
"step": 97
},
{
"epoch": 1.0025575447570332,
"grad_norm": 1.7155461125801181,
"learning_rate": 0.00015506527777142453,
"loss": 1.4022,
"step": 98
},
{
"epoch": 1.0127877237851663,
"grad_norm": 1.0329131057981322,
"learning_rate": 0.00015486401032164434,
"loss": 1.0962,
"step": 99
},
{
"epoch": 1.0230179028132993,
"grad_norm": 1.5658900134919402,
"learning_rate": 0.00015465885602012117,
"loss": 1.1252,
"step": 100
},
{
"epoch": 1.0332480818414322,
"grad_norm": 0.8015741030349912,
"learning_rate": 0.00015444982551822604,
"loss": 1.1044,
"step": 101
},
{
"epoch": 1.0434782608695652,
"grad_norm": 1.3138293387151978,
"learning_rate": 0.00015423692966857788,
"loss": 1.1138,
"step": 102
},
{
"epoch": 1.0537084398976981,
"grad_norm": 0.8910062603103389,
"learning_rate": 0.00015402017952447983,
"loss": 1.0804,
"step": 103
},
{
"epoch": 1.0639386189258313,
"grad_norm": 1.4375718907843533,
"learning_rate": 0.00015379958633934555,
"loss": 1.1212,
"step": 104
},
{
"epoch": 1.0741687979539642,
"grad_norm": 1.2721636337229971,
"learning_rate": 0.0001535751615661149,
"loss": 1.1,
"step": 105
},
{
"epoch": 1.0843989769820972,
"grad_norm": 1.0105810629121363,
"learning_rate": 0.00015334691685665928,
"loss": 1.096,
"step": 106
},
{
"epoch": 1.0946291560102301,
"grad_norm": 1.0900734221547557,
"learning_rate": 0.00015311486406117668,
"loss": 1.0882,
"step": 107
},
{
"epoch": 1.104859335038363,
"grad_norm": 1.1336612336243286,
"learning_rate": 0.00015287901522757652,
"loss": 1.1214,
"step": 108
},
{
"epoch": 1.1150895140664963,
"grad_norm": 1.1569430105162481,
"learning_rate": 0.000152639382600854,
"loss": 1.0963,
"step": 109
},
{
"epoch": 1.1253196930946292,
"grad_norm": 1.1569271586356937,
"learning_rate": 0.00015239597862245452,
"loss": 1.0855,
"step": 110
},
{
"epoch": 1.1355498721227621,
"grad_norm": 1.137534239049028,
"learning_rate": 0.00015214881592962753,
"loss": 1.094,
"step": 111
},
{
"epoch": 1.145780051150895,
"grad_norm": 1.2673351118799951,
"learning_rate": 0.00015189790735477062,
"loss": 1.0819,
"step": 112
},
{
"epoch": 1.156010230179028,
"grad_norm": 0.8230868750468249,
"learning_rate": 0.00015164326592476316,
"loss": 1.099,
"step": 113
},
{
"epoch": 1.1662404092071612,
"grad_norm": 1.0550234081618097,
"learning_rate": 0.00015138490486028998,
"loss": 1.0887,
"step": 114
},
{
"epoch": 1.1764705882352942,
"grad_norm": 1.2808691339484497,
"learning_rate": 0.000151122837575155,
"loss": 1.0879,
"step": 115
},
{
"epoch": 1.186700767263427,
"grad_norm": 1.0207057907200907,
"learning_rate": 0.00015085707767558475,
"loss": 1.0974,
"step": 116
},
{
"epoch": 1.19693094629156,
"grad_norm": 1.7224394340286069,
"learning_rate": 0.00015058763895952194,
"loss": 1.1016,
"step": 117
},
{
"epoch": 1.207161125319693,
"grad_norm": 0.5474668831444041,
"learning_rate": 0.00015031453541590925,
"loss": 1.0789,
"step": 118
},
{
"epoch": 1.2173913043478262,
"grad_norm": 1.2675844883943275,
"learning_rate": 0.00015003778122396277,
"loss": 1.0851,
"step": 119
},
{
"epoch": 1.227621483375959,
"grad_norm": 1.2798593237994211,
"learning_rate": 0.0001497573907524361,
"loss": 1.1047,
"step": 120
},
{
"epoch": 1.237851662404092,
"grad_norm": 1.1502336899283057,
"learning_rate": 0.00014947337855887406,
"loss": 1.0943,
"step": 121
},
{
"epoch": 1.248081841432225,
"grad_norm": 1.2423026547868323,
"learning_rate": 0.00014918575938885725,
"loss": 1.0896,
"step": 122
},
{
"epoch": 1.258312020460358,
"grad_norm": 0.8127468643676744,
"learning_rate": 0.00014889454817523608,
"loss": 1.0984,
"step": 123
},
{
"epoch": 1.2685421994884911,
"grad_norm": 1.1236998604735062,
"learning_rate": 0.00014859976003735572,
"loss": 1.091,
"step": 124
},
{
"epoch": 1.278772378516624,
"grad_norm": 1.0625241691652207,
"learning_rate": 0.0001483014102802711,
"loss": 1.0749,
"step": 125
},
{
"epoch": 1.289002557544757,
"grad_norm": 1.2427086503303664,
"learning_rate": 0.00014799951439395221,
"loss": 1.0901,
"step": 126
},
{
"epoch": 1.29923273657289,
"grad_norm": 0.9886108496701173,
"learning_rate": 0.00014769408805247986,
"loss": 1.0848,
"step": 127
},
{
"epoch": 1.309462915601023,
"grad_norm": 1.017219160331291,
"learning_rate": 0.0001473851471132321,
"loss": 1.0897,
"step": 128
},
{
"epoch": 1.319693094629156,
"grad_norm": 1.4542703012009928,
"learning_rate": 0.00014707270761606063,
"loss": 1.0695,
"step": 129
},
{
"epoch": 1.329923273657289,
"grad_norm": 1.0782706459189146,
"learning_rate": 0.00014675678578245828,
"loss": 1.0895,
"step": 130
},
{
"epoch": 1.340153452685422,
"grad_norm": 1.0968316967604133,
"learning_rate": 0.00014643739801471667,
"loss": 1.1003,
"step": 131
},
{
"epoch": 1.350383631713555,
"grad_norm": 0.9003561398004416,
"learning_rate": 0.00014611456089507464,
"loss": 1.098,
"step": 132
},
{
"epoch": 1.3606138107416879,
"grad_norm": 1.351034619458769,
"learning_rate": 0.00014578829118485742,
"loss": 1.0698,
"step": 133
},
{
"epoch": 1.370843989769821,
"grad_norm": 0.8658183503538551,
"learning_rate": 0.00014545860582360624,
"loss": 1.1071,
"step": 134
},
{
"epoch": 1.381074168797954,
"grad_norm": 1.0177123952310823,
"learning_rate": 0.00014512552192819897,
"loss": 1.0869,
"step": 135
},
{
"epoch": 1.391304347826087,
"grad_norm": 1.4195927565971158,
"learning_rate": 0.0001447890567919614,
"loss": 1.0954,
"step": 136
},
{
"epoch": 1.40153452685422,
"grad_norm": 0.9931491307968467,
"learning_rate": 0.00014444922788376934,
"loss": 1.0784,
"step": 137
},
{
"epoch": 1.4117647058823528,
"grad_norm": 1.270639141634098,
"learning_rate": 0.00014410605284714175,
"loss": 1.0888,
"step": 138
},
{
"epoch": 1.421994884910486,
"grad_norm": 0.8408471380854269,
"learning_rate": 0.0001437595494993246,
"loss": 1.0842,
"step": 139
},
{
"epoch": 1.432225063938619,
"grad_norm": 1.380164251857079,
"learning_rate": 0.000143409735830366,
"loss": 1.0795,
"step": 140
},
{
"epoch": 1.4424552429667519,
"grad_norm": 0.9896679635139878,
"learning_rate": 0.00014305663000218193,
"loss": 1.0907,
"step": 141
},
{
"epoch": 1.452685421994885,
"grad_norm": 1.412072301199199,
"learning_rate": 0.00014270025034761352,
"loss": 1.0817,
"step": 142
},
{
"epoch": 1.4629156010230178,
"grad_norm": 0.770940428011719,
"learning_rate": 0.000142340615369475,
"loss": 1.0819,
"step": 143
},
{
"epoch": 1.473145780051151,
"grad_norm": 0.7983918736084521,
"learning_rate": 0.00014197774373959327,
"loss": 1.0931,
"step": 144
},
{
"epoch": 1.4833759590792839,
"grad_norm": 1.3538690920598053,
"learning_rate": 0.00014161165429783844,
"loss": 1.0884,
"step": 145
},
{
"epoch": 1.4936061381074168,
"grad_norm": 0.8169757443568035,
"learning_rate": 0.0001412423660511456,
"loss": 1.0924,
"step": 146
},
{
"epoch": 1.50383631713555,
"grad_norm": 0.9256987946838048,
"learning_rate": 0.00014086989817252803,
"loss": 1.0785,
"step": 147
},
{
"epoch": 1.5140664961636827,
"grad_norm": 1.1631633494660514,
"learning_rate": 0.00014049427000008185,
"loss": 1.0699,
"step": 148
},
{
"epoch": 1.5242966751918159,
"grad_norm": 1.212333243880141,
"learning_rate": 0.00014011550103598176,
"loss": 1.064,
"step": 149
},
{
"epoch": 1.5345268542199488,
"grad_norm": 0.9967182911949714,
"learning_rate": 0.0001397336109454689,
"loss": 1.1002,
"step": 150
},
{
"epoch": 1.5447570332480818,
"grad_norm": 1.1502313981018193,
"learning_rate": 0.0001393486195558295,
"loss": 1.0709,
"step": 151
},
{
"epoch": 1.554987212276215,
"grad_norm": 1.311911957474106,
"learning_rate": 0.00013896054685536566,
"loss": 1.0717,
"step": 152
},
{
"epoch": 1.5652173913043477,
"grad_norm": 0.8141917822120857,
"learning_rate": 0.00013856941299235752,
"loss": 1.0714,
"step": 153
},
{
"epoch": 1.5754475703324808,
"grad_norm": 0.5826303398204319,
"learning_rate": 0.00013817523827401715,
"loss": 1.0825,
"step": 154
},
{
"epoch": 1.5856777493606138,
"grad_norm": 1.1035174246946824,
"learning_rate": 0.00013777804316543438,
"loss": 1.0583,
"step": 155
},
{
"epoch": 1.5959079283887467,
"grad_norm": 1.483439389847773,
"learning_rate": 0.00013737784828851405,
"loss": 1.0998,
"step": 156
},
{
"epoch": 1.60613810741688,
"grad_norm": 0.5590063439764236,
"learning_rate": 0.0001369746744209055,
"loss": 1.0814,
"step": 157
},
{
"epoch": 1.6163682864450126,
"grad_norm": 1.427302602685382,
"learning_rate": 0.00013656854249492382,
"loss": 1.0672,
"step": 158
},
{
"epoch": 1.6265984654731458,
"grad_norm": 0.9907320262762309,
"learning_rate": 0.00013615947359646295,
"loss": 1.077,
"step": 159
},
{
"epoch": 1.6368286445012787,
"grad_norm": 1.3253726577143705,
"learning_rate": 0.00013574748896390105,
"loss": 1.0831,
"step": 160
},
{
"epoch": 1.6470588235294117,
"grad_norm": 0.6170129994977172,
"learning_rate": 0.00013533260998699776,
"loss": 1.0808,
"step": 161
},
{
"epoch": 1.6572890025575449,
"grad_norm": 0.8974715249718852,
"learning_rate": 0.00013491485820578373,
"loss": 1.0609,
"step": 162
},
{
"epoch": 1.6675191815856778,
"grad_norm": 0.8218782327367921,
"learning_rate": 0.00013449425530944218,
"loss": 1.0822,
"step": 163
},
{
"epoch": 1.6777493606138107,
"grad_norm": 0.6783603661700464,
"learning_rate": 0.00013407082313518292,
"loss": 1.0771,
"step": 164
},
{
"epoch": 1.6879795396419437,
"grad_norm": 0.6976792612146404,
"learning_rate": 0.0001336445836671086,
"loss": 1.0853,
"step": 165
},
{
"epoch": 1.6982097186700766,
"grad_norm": 0.823428019130252,
"learning_rate": 0.0001332155590350732,
"loss": 1.0838,
"step": 166
},
{
"epoch": 1.7084398976982098,
"grad_norm": 1.0953340521038164,
"learning_rate": 0.0001327837715135332,
"loss": 1.082,
"step": 167
},
{
"epoch": 1.7186700767263428,
"grad_norm": 0.9474195557921864,
"learning_rate": 0.00013234924352039103,
"loss": 1.0802,
"step": 168
},
{
"epoch": 1.7289002557544757,
"grad_norm": 0.9444470757389084,
"learning_rate": 0.00013191199761583124,
"loss": 1.0887,
"step": 169
},
{
"epoch": 1.7391304347826086,
"grad_norm": 1.101212086185109,
"learning_rate": 0.00013147205650114913,
"loss": 1.0718,
"step": 170
},
{
"epoch": 1.7493606138107416,
"grad_norm": 1.0553531873256305,
"learning_rate": 0.0001310294430175722,
"loss": 1.0788,
"step": 171
},
{
"epoch": 1.7595907928388748,
"grad_norm": 1.05582162764911,
"learning_rate": 0.00013058418014507412,
"loss": 1.0879,
"step": 172
},
{
"epoch": 1.7698209718670077,
"grad_norm": 0.9944536264830629,
"learning_rate": 0.00013013629100118183,
"loss": 1.0721,
"step": 173
},
{
"epoch": 1.7800511508951407,
"grad_norm": 1.1307659477843484,
"learning_rate": 0.00012968579883977508,
"loss": 1.0737,
"step": 174
},
{
"epoch": 1.7902813299232738,
"grad_norm": 1.2420812191214468,
"learning_rate": 0.00012923272704987943,
"loss": 1.0742,
"step": 175
},
{
"epoch": 1.8005115089514065,
"grad_norm": 1.0472853919924061,
"learning_rate": 0.00012877709915445155,
"loss": 1.0721,
"step": 176
},
{
"epoch": 1.8107416879795397,
"grad_norm": 0.6947884508219546,
"learning_rate": 0.00012831893880915822,
"loss": 1.0555,
"step": 177
},
{
"epoch": 1.8209718670076727,
"grad_norm": 0.7757404896202875,
"learning_rate": 0.00012785826980114798,
"loss": 1.0804,
"step": 178
},
{
"epoch": 1.8312020460358056,
"grad_norm": 1.253778618594067,
"learning_rate": 0.0001273951160478163,
"loss": 1.063,
"step": 179
},
{
"epoch": 1.8414322250639388,
"grad_norm": 0.5979958514242996,
"learning_rate": 0.00012692950159556358,
"loss": 1.0666,
"step": 180
},
{
"epoch": 1.8516624040920715,
"grad_norm": 0.8662747748641294,
"learning_rate": 0.00012646145061854697,
"loss": 1.0703,
"step": 181
},
{
"epoch": 1.8618925831202047,
"grad_norm": 0.8287006788806506,
"learning_rate": 0.00012599098741742504,
"loss": 1.0571,
"step": 182
},
{
"epoch": 1.8721227621483376,
"grad_norm": 1.0265713289334373,
"learning_rate": 0.00012551813641809622,
"loss": 1.0706,
"step": 183
},
{
"epoch": 1.8823529411764706,
"grad_norm": 1.2306509484226458,
"learning_rate": 0.0001250429221704306,
"loss": 1.0779,
"step": 184
},
{
"epoch": 1.8925831202046037,
"grad_norm": 0.8951251773522786,
"learning_rate": 0.00012456536934699552,
"loss": 1.064,
"step": 185
},
{
"epoch": 1.9028132992327365,
"grad_norm": 1.1181307738147266,
"learning_rate": 0.0001240855027417742,
"loss": 1.0585,
"step": 186
},
{
"epoch": 1.9130434782608696,
"grad_norm": 0.977542614018667,
"learning_rate": 0.00012360334726887887,
"loss": 1.0672,
"step": 187
},
{
"epoch": 1.9232736572890026,
"grad_norm": 1.1637804133002638,
"learning_rate": 0.00012311892796125704,
"loss": 1.0713,
"step": 188
},
{
"epoch": 1.9335038363171355,
"grad_norm": 0.445419436951453,
"learning_rate": 0.0001226322699693918,
"loss": 1.0536,
"step": 189
},
{
"epoch": 1.9437340153452687,
"grad_norm": 0.9034412973205753,
"learning_rate": 0.00012214339855999624,
"loss": 1.0807,
"step": 190
},
{
"epoch": 1.9539641943734014,
"grad_norm": 0.5236086539347802,
"learning_rate": 0.00012165233911470136,
"loss": 1.0777,
"step": 191
},
{
"epoch": 1.9641943734015346,
"grad_norm": 0.6745807390622581,
"learning_rate": 0.00012115911712873851,
"loss": 1.0525,
"step": 192
},
{
"epoch": 1.9744245524296675,
"grad_norm": 0.5551428173478703,
"learning_rate": 0.00012066375820961558,
"loss": 1.0617,
"step": 193
},
{
"epoch": 1.9846547314578005,
"grad_norm": 0.5930456521809883,
"learning_rate": 0.00012016628807578756,
"loss": 1.0682,
"step": 194
},
{
"epoch": 1.9948849104859336,
"grad_norm": 0.5333039982670191,
"learning_rate": 0.00011966673255532119,
"loss": 1.1518,
"step": 195
},
{
"epoch": 2.0051150895140664,
"grad_norm": 0.7368227183464856,
"learning_rate": 0.00011916511758455407,
"loss": 1.226,
"step": 196
},
{
"epoch": 2.0153452685421995,
"grad_norm": 1.0538129513540742,
"learning_rate": 0.00011866146920674807,
"loss": 1.068,
"step": 197
},
{
"epoch": 2.0255754475703327,
"grad_norm": 1.2094094160759923,
"learning_rate": 0.0001181558135707371,
"loss": 1.0502,
"step": 198
},
{
"epoch": 2.0358056265984654,
"grad_norm": 0.915224534709575,
"learning_rate": 0.00011764817692956966,
"loss": 1.0286,
"step": 199
},
{
"epoch": 2.0460358056265986,
"grad_norm": 1.0711022599910567,
"learning_rate": 0.00011713858563914562,
"loss": 1.0747,
"step": 200
},
{
"epoch": 2.0562659846547313,
"grad_norm": 1.0787442473268498,
"learning_rate": 0.00011662706615684803,
"loss": 1.045,
"step": 201
},
{
"epoch": 2.0664961636828645,
"grad_norm": 0.8778392977870085,
"learning_rate": 0.00011611364504016935,
"loss": 1.0678,
"step": 202
},
{
"epoch": 2.0767263427109977,
"grad_norm": 0.9910713045345557,
"learning_rate": 0.00011559834894533275,
"loss": 1.0458,
"step": 203
},
{
"epoch": 2.0869565217391304,
"grad_norm": 1.1592891619781418,
"learning_rate": 0.00011508120462590794,
"loss": 1.0461,
"step": 204
},
{
"epoch": 2.0971867007672635,
"grad_norm": 0.6931145505443533,
"learning_rate": 0.00011456223893142238,
"loss": 1.0407,
"step": 205
},
{
"epoch": 2.1074168797953963,
"grad_norm": 0.8009294768801537,
"learning_rate": 0.0001140414788059672,
"loss": 1.0534,
"step": 206
},
{
"epoch": 2.1176470588235294,
"grad_norm": 0.6527102431957467,
"learning_rate": 0.00011351895128679823,
"loss": 1.0577,
"step": 207
},
{
"epoch": 2.1278772378516626,
"grad_norm": 0.5808810881566437,
"learning_rate": 0.00011299468350293232,
"loss": 1.0592,
"step": 208
},
{
"epoch": 2.1381074168797953,
"grad_norm": 0.8896635325022703,
"learning_rate": 0.00011246870267373885,
"loss": 1.069,
"step": 209
},
{
"epoch": 2.1483375959079285,
"grad_norm": 1.0873242378114067,
"learning_rate": 0.00011194103610752655,
"loss": 1.0454,
"step": 210
},
{
"epoch": 2.1585677749360612,
"grad_norm": 0.9489754319671625,
"learning_rate": 0.00011141171120012552,
"loss": 1.0723,
"step": 211
},
{
"epoch": 2.1687979539641944,
"grad_norm": 0.8030522232247485,
"learning_rate": 0.0001108807554334651,
"loss": 1.0428,
"step": 212
},
{
"epoch": 2.1790281329923276,
"grad_norm": 0.5529249514757153,
"learning_rate": 0.00011034819637414686,
"loss": 1.061,
"step": 213
},
{
"epoch": 2.1892583120204603,
"grad_norm": 0.4631583007632592,
"learning_rate": 0.00010981406167201354,
"loss": 1.0355,
"step": 214
},
{
"epoch": 2.1994884910485935,
"grad_norm": 0.4099197978603843,
"learning_rate": 0.0001092783790587133,
"loss": 1.0777,
"step": 215
},
{
"epoch": 2.209718670076726,
"grad_norm": 0.4121757237952397,
"learning_rate": 0.00010874117634626011,
"loss": 1.0541,
"step": 216
},
{
"epoch": 2.2199488491048593,
"grad_norm": 0.4349846280052258,
"learning_rate": 0.00010820248142558965,
"loss": 1.0435,
"step": 217
},
{
"epoch": 2.2301790281329925,
"grad_norm": 0.3960108429870153,
"learning_rate": 0.00010766232226511142,
"loss": 1.0513,
"step": 218
},
{
"epoch": 2.2404092071611252,
"grad_norm": 0.4522590868845693,
"learning_rate": 0.00010712072690925638,
"loss": 1.0509,
"step": 219
},
{
"epoch": 2.2506393861892584,
"grad_norm": 0.4862701100635931,
"learning_rate": 0.00010657772347702118,
"loss": 1.0325,
"step": 220
},
{
"epoch": 2.260869565217391,
"grad_norm": 0.44763620371165586,
"learning_rate": 0.00010603334016050808,
"loss": 1.0369,
"step": 221
},
{
"epoch": 2.2710997442455243,
"grad_norm": 0.5580650650662916,
"learning_rate": 0.00010548760522346138,
"loss": 1.0414,
"step": 222
},
{
"epoch": 2.2813299232736575,
"grad_norm": 0.6352266377825914,
"learning_rate": 0.00010494054699979992,
"loss": 1.056,
"step": 223
},
{
"epoch": 2.29156010230179,
"grad_norm": 0.617790095766774,
"learning_rate": 0.00010439219389214595,
"loss": 1.0573,
"step": 224
},
{
"epoch": 2.3017902813299234,
"grad_norm": 0.6417970276744828,
"learning_rate": 0.0001038425743703507,
"loss": 1.0412,
"step": 225
},
{
"epoch": 2.312020460358056,
"grad_norm": 0.7082597904375986,
"learning_rate": 0.00010329171697001608,
"loss": 1.0366,
"step": 226
},
{
"epoch": 2.3222506393861893,
"grad_norm": 0.8496266841221832,
"learning_rate": 0.0001027396502910132,
"loss": 1.0451,
"step": 227
},
{
"epoch": 2.3324808184143224,
"grad_norm": 1.0308330001421908,
"learning_rate": 0.0001021864029959975,
"loss": 1.0428,
"step": 228
},
{
"epoch": 2.342710997442455,
"grad_norm": 1.026355847644825,
"learning_rate": 0.00010163200380892063,
"loss": 1.0612,
"step": 229
},
{
"epoch": 2.3529411764705883,
"grad_norm": 0.7560985076254375,
"learning_rate": 0.00010107648151353916,
"loss": 1.0247,
"step": 230
},
{
"epoch": 2.363171355498721,
"grad_norm": 0.41561156978885305,
"learning_rate": 0.00010051986495192008,
"loss": 1.0363,
"step": 231
},
{
"epoch": 2.373401534526854,
"grad_norm": 0.38675030483907014,
"learning_rate": 9.99621830229434e-05,
"loss": 1.05,
"step": 232
},
{
"epoch": 2.3836317135549874,
"grad_norm": 0.4804093390079753,
"learning_rate": 9.94034646808018e-05,
"loss": 1.0537,
"step": 233
},
{
"epoch": 2.39386189258312,
"grad_norm": 0.5877095225822028,
"learning_rate": 9.884373893349725e-05,
"loss": 1.0273,
"step": 234
},
{
"epoch": 2.4040920716112533,
"grad_norm": 0.6010921886045744,
"learning_rate": 9.828303484133515e-05,
"loss": 1.053,
"step": 235
},
{
"epoch": 2.414322250639386,
"grad_norm": 0.4918851450838751,
"learning_rate": 9.772138151541522e-05,
"loss": 1.0364,
"step": 236
},
{
"epoch": 2.424552429667519,
"grad_norm": 0.3407705594769831,
"learning_rate": 9.715880811612044e-05,
"loss": 1.0331,
"step": 237
},
{
"epoch": 2.4347826086956523,
"grad_norm": 0.2468632428242675,
"learning_rate": 9.659534385160289e-05,
"loss": 1.0323,
"step": 238
},
{
"epoch": 2.445012787723785,
"grad_norm": 0.26175126594915626,
"learning_rate": 9.603101797626729e-05,
"loss": 1.0491,
"step": 239
},
{
"epoch": 2.455242966751918,
"grad_norm": 0.3250566758468044,
"learning_rate": 9.546585978925221e-05,
"loss": 1.0127,
"step": 240
},
{
"epoch": 2.4654731457800514,
"grad_norm": 0.3733853550848411,
"learning_rate": 9.489989863290885e-05,
"loss": 1.0637,
"step": 241
},
{
"epoch": 2.475703324808184,
"grad_norm": 0.45748627752418863,
"learning_rate": 9.433316389127768e-05,
"loss": 1.038,
"step": 242
},
{
"epoch": 2.4859335038363173,
"grad_norm": 0.432153145698656,
"learning_rate": 9.37656849885628e-05,
"loss": 1.0441,
"step": 243
},
{
"epoch": 2.49616368286445,
"grad_norm": 0.3361885905419595,
"learning_rate": 9.319749138760424e-05,
"loss": 1.0317,
"step": 244
},
{
"epoch": 2.506393861892583,
"grad_norm": 0.26392084893370804,
"learning_rate": 9.262861258834833e-05,
"loss": 1.0353,
"step": 245
},
{
"epoch": 2.516624040920716,
"grad_norm": 0.25278206548366783,
"learning_rate": 9.205907812631616e-05,
"loss": 1.0211,
"step": 246
},
{
"epoch": 2.526854219948849,
"grad_norm": 0.2853155639467533,
"learning_rate": 9.148891757106999e-05,
"loss": 1.0381,
"step": 247
},
{
"epoch": 2.5370843989769822,
"grad_norm": 0.28133618292890095,
"learning_rate": 9.091816052467817e-05,
"loss": 1.045,
"step": 248
},
{
"epoch": 2.547314578005115,
"grad_norm": 0.26415138394214893,
"learning_rate": 9.034683662017812e-05,
"loss": 1.0339,
"step": 249
},
{
"epoch": 2.557544757033248,
"grad_norm": 0.23100116731609785,
"learning_rate": 8.977497552003785e-05,
"loss": 1.0297,
"step": 250
},
{
"epoch": 2.5677749360613813,
"grad_norm": 0.25137211273391913,
"learning_rate": 8.920260691461602e-05,
"loss": 1.0474,
"step": 251
},
{
"epoch": 2.578005115089514,
"grad_norm": 0.25925530222353527,
"learning_rate": 8.862976052062034e-05,
"loss": 1.0478,
"step": 252
},
{
"epoch": 2.588235294117647,
"grad_norm": 0.21181713461857493,
"learning_rate": 8.805646607956467e-05,
"loss": 1.0384,
"step": 253
},
{
"epoch": 2.59846547314578,
"grad_norm": 0.20383781215036986,
"learning_rate": 8.748275335622506e-05,
"loss": 1.0352,
"step": 254
},
{
"epoch": 2.608695652173913,
"grad_norm": 0.22216426938201625,
"learning_rate": 8.69086521370942e-05,
"loss": 1.0251,
"step": 255
},
{
"epoch": 2.618925831202046,
"grad_norm": 0.2339550911616179,
"learning_rate": 8.633419222883508e-05,
"loss": 1.0388,
"step": 256
},
{
"epoch": 2.629156010230179,
"grad_norm": 0.29608330426201757,
"learning_rate": 8.575940345673337e-05,
"loss": 1.0415,
"step": 257
},
{
"epoch": 2.639386189258312,
"grad_norm": 0.282754017428277,
"learning_rate": 8.518431566314901e-05,
"loss": 1.0338,
"step": 258
},
{
"epoch": 2.649616368286445,
"grad_norm": 0.21590808755680316,
"learning_rate": 8.460895870596675e-05,
"loss": 1.0455,
"step": 259
},
{
"epoch": 2.659846547314578,
"grad_norm": 0.22038275314532527,
"learning_rate": 8.4033362457046e-05,
"loss": 1.0446,
"step": 260
},
{
"epoch": 2.670076726342711,
"grad_norm": 0.19903936327716265,
"learning_rate": 8.345755680066993e-05,
"loss": 1.0282,
"step": 261
},
{
"epoch": 2.680306905370844,
"grad_norm": 0.17558239320808622,
"learning_rate": 8.288157163199389e-05,
"loss": 1.0278,
"step": 262
},
{
"epoch": 2.690537084398977,
"grad_norm": 0.20248735135033116,
"learning_rate": 8.230543685549333e-05,
"loss": 1.0317,
"step": 263
},
{
"epoch": 2.70076726342711,
"grad_norm": 0.23742588094542533,
"learning_rate": 8.17291823834111e-05,
"loss": 1.0326,
"step": 264
},
{
"epoch": 2.710997442455243,
"grad_norm": 0.3032160882786812,
"learning_rate": 8.115283813420459e-05,
"loss": 1.0375,
"step": 265
},
{
"epoch": 2.7212276214833757,
"grad_norm": 0.2962448679587415,
"learning_rate": 8.057643403099221e-05,
"loss": 1.0584,
"step": 266
},
{
"epoch": 2.731457800511509,
"grad_norm": 0.30722625928358643,
"learning_rate": 8e-05,
"loss": 1.0395,
"step": 267
},
{
"epoch": 2.741687979539642,
"grad_norm": 0.28307960709841534,
"learning_rate": 7.94235659690078e-05,
"loss": 1.0369,
"step": 268
},
{
"epoch": 2.7519181585677748,
"grad_norm": 0.21321155330317557,
"learning_rate": 7.884716186579545e-05,
"loss": 1.0532,
"step": 269
},
{
"epoch": 2.762148337595908,
"grad_norm": 0.2443283053996415,
"learning_rate": 7.827081761658892e-05,
"loss": 1.0266,
"step": 270
},
{
"epoch": 2.772378516624041,
"grad_norm": 0.2904460725842854,
"learning_rate": 7.76945631445067e-05,
"loss": 1.0344,
"step": 271
},
{
"epoch": 2.782608695652174,
"grad_norm": 0.25431043958852784,
"learning_rate": 7.711842836800614e-05,
"loss": 1.0285,
"step": 272
},
{
"epoch": 2.792838874680307,
"grad_norm": 0.25586265632907745,
"learning_rate": 7.654244319933009e-05,
"loss": 1.0272,
"step": 273
},
{
"epoch": 2.80306905370844,
"grad_norm": 0.23485094953105404,
"learning_rate": 7.596663754295404e-05,
"loss": 1.0427,
"step": 274
},
{
"epoch": 2.813299232736573,
"grad_norm": 0.17957023788842033,
"learning_rate": 7.539104129403327e-05,
"loss": 1.0474,
"step": 275
},
{
"epoch": 2.8235294117647056,
"grad_norm": 0.160508230528404,
"learning_rate": 7.4815684336851e-05,
"loss": 1.0445,
"step": 276
},
{
"epoch": 2.833759590792839,
"grad_norm": 0.21356478024874126,
"learning_rate": 7.424059654326664e-05,
"loss": 1.04,
"step": 277
},
{
"epoch": 2.843989769820972,
"grad_norm": 0.21888445256566497,
"learning_rate": 7.366580777116495e-05,
"loss": 1.0406,
"step": 278
},
{
"epoch": 2.8542199488491047,
"grad_norm": 0.2041365287775187,
"learning_rate": 7.309134786290583e-05,
"loss": 1.0321,
"step": 279
},
{
"epoch": 2.864450127877238,
"grad_norm": 0.19151321727068246,
"learning_rate": 7.251724664377497e-05,
"loss": 1.0371,
"step": 280
},
{
"epoch": 2.874680306905371,
"grad_norm": 0.18344693741146628,
"learning_rate": 7.194353392043534e-05,
"loss": 1.039,
"step": 281
},
{
"epoch": 2.8849104859335037,
"grad_norm": 0.19330950475902522,
"learning_rate": 7.13702394793797e-05,
"loss": 1.0364,
"step": 282
},
{
"epoch": 2.895140664961637,
"grad_norm": 0.1990648393315987,
"learning_rate": 7.079739308538399e-05,
"loss": 1.0277,
"step": 283
},
{
"epoch": 2.90537084398977,
"grad_norm": 0.2262885121158685,
"learning_rate": 7.022502447996215e-05,
"loss": 1.0275,
"step": 284
},
{
"epoch": 2.915601023017903,
"grad_norm": 0.20233291351840074,
"learning_rate": 6.965316337982191e-05,
"loss": 1.0381,
"step": 285
},
{
"epoch": 2.9258312020460355,
"grad_norm": 0.19513363820566396,
"learning_rate": 6.908183947532184e-05,
"loss": 1.0342,
"step": 286
},
{
"epoch": 2.9360613810741687,
"grad_norm": 0.20380750046128057,
"learning_rate": 6.851108242893002e-05,
"loss": 1.0377,
"step": 287
},
{
"epoch": 2.946291560102302,
"grad_norm": 0.1866318852041668,
"learning_rate": 6.794092187368387e-05,
"loss": 1.0428,
"step": 288
},
{
"epoch": 2.9565217391304346,
"grad_norm": 0.15744510473495013,
"learning_rate": 6.737138741165168e-05,
"loss": 1.0503,
"step": 289
},
{
"epoch": 2.9667519181585678,
"grad_norm": 0.1652822650571465,
"learning_rate": 6.680250861239581e-05,
"loss": 1.035,
"step": 290
},
{
"epoch": 2.976982097186701,
"grad_norm": 0.15546583646318948,
"learning_rate": 6.623431501143723e-05,
"loss": 1.0313,
"step": 291
},
{
"epoch": 2.9872122762148337,
"grad_norm": 0.29006532173297445,
"learning_rate": 6.566683610872231e-05,
"loss": 1.0564,
"step": 292
},
{
"epoch": 2.997442455242967,
"grad_norm": 0.2268887255063016,
"learning_rate": 6.510010136709118e-05,
"loss": 1.2037,
"step": 293
},
{
"epoch": 3.0076726342710995,
"grad_norm": 0.2539730951218038,
"learning_rate": 6.453414021074781e-05,
"loss": 1.1394,
"step": 294
},
{
"epoch": 3.0179028132992327,
"grad_norm": 0.2231294668730024,
"learning_rate": 6.396898202373277e-05,
"loss": 1.0223,
"step": 295
},
{
"epoch": 3.028132992327366,
"grad_norm": 0.20401904180469313,
"learning_rate": 6.340465614839714e-05,
"loss": 1.0336,
"step": 296
},
{
"epoch": 3.0383631713554986,
"grad_norm": 0.18979057876237626,
"learning_rate": 6.284119188387957e-05,
"loss": 1.0107,
"step": 297
},
{
"epoch": 3.0485933503836318,
"grad_norm": 0.16567071328991922,
"learning_rate": 6.227861848458481e-05,
"loss": 1.0134,
"step": 298
},
{
"epoch": 3.0588235294117645,
"grad_norm": 0.2788005001606031,
"learning_rate": 6.171696515866488e-05,
"loss": 1.0289,
"step": 299
},
{
"epoch": 3.0690537084398977,
"grad_norm": 0.13520429300745568,
"learning_rate": 6.115626106650273e-05,
"loss": 1.0297,
"step": 300
},
{
"epoch": 3.079283887468031,
"grad_norm": 0.21854013343576806,
"learning_rate": 6.059653531919823e-05,
"loss": 1.0282,
"step": 301
},
{
"epoch": 3.0895140664961636,
"grad_norm": 0.1818832819994029,
"learning_rate": 6.0037816977056625e-05,
"loss": 1.0531,
"step": 302
},
{
"epoch": 3.0997442455242967,
"grad_norm": 0.18589522528315505,
"learning_rate": 5.9480135048079964e-05,
"loss": 1.0113,
"step": 303
},
{
"epoch": 3.10997442455243,
"grad_norm": 0.18063564176637392,
"learning_rate": 5.892351848646087e-05,
"loss": 1.0394,
"step": 304
},
{
"epoch": 3.1202046035805626,
"grad_norm": 0.147586744880159,
"learning_rate": 5.836799619107937e-05,
"loss": 1.0365,
"step": 305
},
{
"epoch": 3.130434782608696,
"grad_norm": 0.17675912154183307,
"learning_rate": 5.781359700400254e-05,
"loss": 1.0039,
"step": 306
},
{
"epoch": 3.1406649616368285,
"grad_norm": 0.16586699917434417,
"learning_rate": 5.726034970898682e-05,
"loss": 1.0243,
"step": 307
},
{
"epoch": 3.1508951406649617,
"grad_norm": 0.17820259338383218,
"learning_rate": 5.670828302998393e-05,
"loss": 1.0314,
"step": 308
},
{
"epoch": 3.1611253196930944,
"grad_norm": 0.15458082339803622,
"learning_rate": 5.6157425629649314e-05,
"loss": 1.0485,
"step": 309
},
{
"epoch": 3.1713554987212276,
"grad_norm": 0.1643571380269719,
"learning_rate": 5.560780610785406e-05,
"loss": 1.018,
"step": 310
},
{
"epoch": 3.1815856777493607,
"grad_norm": 0.156856618068632,
"learning_rate": 5.5059453000200125e-05,
"loss": 1.0061,
"step": 311
},
{
"epoch": 3.1918158567774935,
"grad_norm": 0.14086641088136265,
"learning_rate": 5.451239477653864e-05,
"loss": 1.0205,
"step": 312
},
{
"epoch": 3.2020460358056266,
"grad_norm": 0.14820087407296167,
"learning_rate": 5.3966659839491936e-05,
"loss": 1.0226,
"step": 313
},
{
"epoch": 3.21227621483376,
"grad_norm": 0.13884477979100748,
"learning_rate": 5.342227652297887e-05,
"loss": 1.001,
"step": 314
},
{
"epoch": 3.2225063938618925,
"grad_norm": 0.14309328916359365,
"learning_rate": 5.287927309074365e-05,
"loss": 1.031,
"step": 315
},
{
"epoch": 3.2327365728900257,
"grad_norm": 0.15252565103321644,
"learning_rate": 5.233767773488859e-05,
"loss": 1.015,
"step": 316
},
{
"epoch": 3.2429667519181584,
"grad_norm": 0.13230625482589622,
"learning_rate": 5.179751857441036e-05,
"loss": 1.0053,
"step": 317
},
{
"epoch": 3.2531969309462916,
"grad_norm": 0.1556826070734683,
"learning_rate": 5.1258823653739914e-05,
"loss": 1.0211,
"step": 318
},
{
"epoch": 3.2634271099744243,
"grad_norm": 0.1371375107565863,
"learning_rate": 5.0721620941286735e-05,
"loss": 1.0143,
"step": 319
},
{
"epoch": 3.2736572890025575,
"grad_norm": 0.13573969073646522,
"learning_rate": 5.018593832798649e-05,
"loss": 1.0375,
"step": 320
},
{
"epoch": 3.2838874680306906,
"grad_norm": 0.11657280744271552,
"learning_rate": 4.965180362585315e-05,
"loss": 1.0253,
"step": 321
},
{
"epoch": 3.2941176470588234,
"grad_norm": 0.11616693511840327,
"learning_rate": 4.911924456653494e-05,
"loss": 1.0209,
"step": 322
},
{
"epoch": 3.3043478260869565,
"grad_norm": 0.12512977851281545,
"learning_rate": 4.8588288799874514e-05,
"loss": 1.0112,
"step": 323
},
{
"epoch": 3.3145780051150897,
"grad_norm": 0.10387393823484337,
"learning_rate": 4.805896389247348e-05,
"loss": 1.0077,
"step": 324
},
{
"epoch": 3.3248081841432224,
"grad_norm": 0.13924261830307932,
"learning_rate": 4.753129732626116e-05,
"loss": 1.0229,
"step": 325
},
{
"epoch": 3.3350383631713556,
"grad_norm": 0.12135011031399273,
"learning_rate": 4.70053164970677e-05,
"loss": 1.0184,
"step": 326
},
{
"epoch": 3.3452685421994883,
"grad_norm": 0.1183989328140461,
"learning_rate": 4.6481048713201825e-05,
"loss": 1.0058,
"step": 327
},
{
"epoch": 3.3554987212276215,
"grad_norm": 0.1284311973823529,
"learning_rate": 4.595852119403282e-05,
"loss": 1.0278,
"step": 328
},
{
"epoch": 3.3657289002557547,
"grad_norm": 0.11430293939389213,
"learning_rate": 4.543776106857765e-05,
"loss": 1.012,
"step": 329
},
{
"epoch": 3.3759590792838874,
"grad_norm": 0.1263289266506516,
"learning_rate": 4.491879537409211e-05,
"loss": 1.0242,
"step": 330
},
{
"epoch": 3.3861892583120206,
"grad_norm": 0.1316142612111793,
"learning_rate": 4.4401651054667274e-05,
"loss": 1.0078,
"step": 331
},
{
"epoch": 3.3964194373401533,
"grad_norm": 0.10490352230994067,
"learning_rate": 4.3886354959830625e-05,
"loss": 1.0141,
"step": 332
},
{
"epoch": 3.4066496163682864,
"grad_norm": 0.1587962681364766,
"learning_rate": 4.3372933843152e-05,
"loss": 1.031,
"step": 333
},
{
"epoch": 3.4168797953964196,
"grad_norm": 0.14806873434228535,
"learning_rate": 4.2861414360854387e-05,
"loss": 1.0261,
"step": 334
},
{
"epoch": 3.4271099744245523,
"grad_norm": 0.12428110117787702,
"learning_rate": 4.2351823070430376e-05,
"loss": 1.0191,
"step": 335
},
{
"epoch": 3.4373401534526855,
"grad_norm": 0.1700524980071232,
"learning_rate": 4.184418642926289e-05,
"loss": 1.0267,
"step": 336
},
{
"epoch": 3.4475703324808182,
"grad_norm": 0.15240023162277883,
"learning_rate": 4.133853079325196e-05,
"loss": 1.025,
"step": 337
},
{
"epoch": 3.4578005115089514,
"grad_norm": 0.11788415451483955,
"learning_rate": 4.083488241544595e-05,
"loss": 1.0459,
"step": 338
},
{
"epoch": 3.4680306905370846,
"grad_norm": 0.16921030624641467,
"learning_rate": 4.033326744467882e-05,
"loss": 1.0112,
"step": 339
},
{
"epoch": 3.4782608695652173,
"grad_norm": 0.11606787523018822,
"learning_rate": 3.983371192421246e-05,
"loss": 1.0306,
"step": 340
},
{
"epoch": 3.4884910485933505,
"grad_norm": 0.1682374876165679,
"learning_rate": 3.933624179038446e-05,
"loss": 1.0185,
"step": 341
},
{
"epoch": 3.498721227621483,
"grad_norm": 0.1565962574632331,
"learning_rate": 3.884088287126151e-05,
"loss": 1.0293,
"step": 342
},
{
"epoch": 3.5089514066496164,
"grad_norm": 0.14324776346066015,
"learning_rate": 3.834766088529867e-05,
"loss": 1.0458,
"step": 343
},
{
"epoch": 3.5191815856777495,
"grad_norm": 0.19529159534261803,
"learning_rate": 3.785660144000378e-05,
"loss": 1.0056,
"step": 344
},
{
"epoch": 3.5294117647058822,
"grad_norm": 0.12801441909483788,
"learning_rate": 3.736773003060821e-05,
"loss": 1.0297,
"step": 345
},
{
"epoch": 3.5396419437340154,
"grad_norm": 0.1503941568635443,
"learning_rate": 3.688107203874301e-05,
"loss": 1.0416,
"step": 346
},
{
"epoch": 3.5498721227621486,
"grad_norm": 0.13361192505091346,
"learning_rate": 3.6396652731121136e-05,
"loss": 1.0204,
"step": 347
},
{
"epoch": 3.5601023017902813,
"grad_norm": 0.12402122906765294,
"learning_rate": 3.5914497258225815e-05,
"loss": 1.0281,
"step": 348
},
{
"epoch": 3.5703324808184145,
"grad_norm": 0.1370545977754261,
"learning_rate": 3.543463065300452e-05,
"loss": 1.0271,
"step": 349
},
{
"epoch": 3.580562659846547,
"grad_norm": 0.10522140449656789,
"learning_rate": 3.49570778295694e-05,
"loss": 1.0253,
"step": 350
},
{
"epoch": 3.5907928388746804,
"grad_norm": 0.13288330038665777,
"learning_rate": 3.448186358190383e-05,
"loss": 1.0155,
"step": 351
},
{
"epoch": 3.601023017902813,
"grad_norm": 0.11850947440760301,
"learning_rate": 3.400901258257501e-05,
"loss": 1.0316,
"step": 352
},
{
"epoch": 3.6112531969309463,
"grad_norm": 0.1207639690615278,
"learning_rate": 3.3538549381453046e-05,
"loss": 1.0147,
"step": 353
},
{
"epoch": 3.6214833759590794,
"grad_norm": 0.10436214175934275,
"learning_rate": 3.307049840443644e-05,
"loss": 1.0158,
"step": 354
},
{
"epoch": 3.631713554987212,
"grad_norm": 0.10939591730050287,
"learning_rate": 3.2604883952183716e-05,
"loss": 1.0219,
"step": 355
},
{
"epoch": 3.6419437340153453,
"grad_norm": 0.10305381344145671,
"learning_rate": 3.214173019885202e-05,
"loss": 1.0165,
"step": 356
},
{
"epoch": 3.6521739130434785,
"grad_norm": 0.10617153493263774,
"learning_rate": 3.1681061190841806e-05,
"loss": 1.0193,
"step": 357
},
{
"epoch": 3.662404092071611,
"grad_norm": 0.1052450380146473,
"learning_rate": 3.122290084554845e-05,
"loss": 1.0309,
"step": 358
},
{
"epoch": 3.6726342710997444,
"grad_norm": 0.09779940640870793,
"learning_rate": 3.076727295012059e-05,
"loss": 1.0106,
"step": 359
},
{
"epoch": 3.682864450127877,
"grad_norm": 0.09865111890866658,
"learning_rate": 3.031420116022493e-05,
"loss": 1.0237,
"step": 360
},
{
"epoch": 3.6930946291560103,
"grad_norm": 0.09761460812236314,
"learning_rate": 2.98637089988182e-05,
"loss": 1.0071,
"step": 361
},
{
"epoch": 3.703324808184143,
"grad_norm": 0.09657195902847425,
"learning_rate": 2.94158198549259e-05,
"loss": 1.0213,
"step": 362
},
{
"epoch": 3.713554987212276,
"grad_norm": 0.10306766572443878,
"learning_rate": 2.8970556982427836e-05,
"loss": 1.0114,
"step": 363
},
{
"epoch": 3.7237851662404093,
"grad_norm": 0.0892256925639781,
"learning_rate": 2.852794349885087e-05,
"loss": 1.0141,
"step": 364
},
{
"epoch": 3.734015345268542,
"grad_norm": 0.0984967903902541,
"learning_rate": 2.8088002384168783e-05,
"loss": 1.0309,
"step": 365
},
{
"epoch": 3.7442455242966752,
"grad_norm": 0.09702833968633048,
"learning_rate": 2.765075647960898e-05,
"loss": 1.0133,
"step": 366
},
{
"epoch": 3.7544757033248084,
"grad_norm": 0.10259443786625837,
"learning_rate": 2.7216228486466856e-05,
"loss": 1.0158,
"step": 367
},
{
"epoch": 3.764705882352941,
"grad_norm": 0.09444360606393558,
"learning_rate": 2.678444096492683e-05,
"loss": 1.02,
"step": 368
},
{
"epoch": 3.7749360613810743,
"grad_norm": 0.11306816830082422,
"learning_rate": 2.6355416332891404e-05,
"loss": 1.0185,
"step": 369
},
{
"epoch": 3.785166240409207,
"grad_norm": 0.09142683583600977,
"learning_rate": 2.592917686481708e-05,
"loss": 1.0038,
"step": 370
},
{
"epoch": 3.79539641943734,
"grad_norm": 0.10949950748671738,
"learning_rate": 2.5505744690557846e-05,
"loss": 1.0376,
"step": 371
},
{
"epoch": 3.805626598465473,
"grad_norm": 0.11343467361764166,
"learning_rate": 2.508514179421629e-05,
"loss": 1.0358,
"step": 372
},
{
"epoch": 3.815856777493606,
"grad_norm": 0.09342791259699781,
"learning_rate": 2.4667390013002254e-05,
"loss": 1.0211,
"step": 373
},
{
"epoch": 3.8260869565217392,
"grad_norm": 0.10858137240897216,
"learning_rate": 2.425251103609898e-05,
"loss": 1.0332,
"step": 374
},
{
"epoch": 3.836317135549872,
"grad_norm": 0.0886186909107238,
"learning_rate": 2.3840526403537095e-05,
"loss": 1.0143,
"step": 375
},
{
"epoch": 3.846547314578005,
"grad_norm": 0.09895029034827527,
"learning_rate": 2.3431457505076205e-05,
"loss": 1.0173,
"step": 376
},
{
"epoch": 3.8567774936061383,
"grad_norm": 0.10546368891288044,
"learning_rate": 2.3025325579094498e-05,
"loss": 1.0362,
"step": 377
},
{
"epoch": 3.867007672634271,
"grad_norm": 0.08194245174545557,
"learning_rate": 2.2622151711485962e-05,
"loss": 1.0124,
"step": 378
},
{
"epoch": 3.877237851662404,
"grad_norm": 0.1114016593112589,
"learning_rate": 2.2221956834565647e-05,
"loss": 1.0139,
"step": 379
},
{
"epoch": 3.887468030690537,
"grad_norm": 0.09816527002580688,
"learning_rate": 2.1824761725982874e-05,
"loss": 1.0523,
"step": 380
},
{
"epoch": 3.89769820971867,
"grad_norm": 0.0799950179310954,
"learning_rate": 2.1430587007642513e-05,
"loss": 0.991,
"step": 381
},
{
"epoch": 3.907928388746803,
"grad_norm": 0.1080492642683735,
"learning_rate": 2.1039453144634364e-05,
"loss": 1.0355,
"step": 382
},
{
"epoch": 3.918158567774936,
"grad_norm": 0.09178099069537385,
"learning_rate": 2.0651380444170527e-05,
"loss": 1.015,
"step": 383
},
{
"epoch": 3.928388746803069,
"grad_norm": 0.09031457518884235,
"learning_rate": 2.026638905453111e-05,
"loss": 1.0229,
"step": 384
},
{
"epoch": 3.938618925831202,
"grad_norm": 0.09836817148107584,
"learning_rate": 1.9884498964018233e-05,
"loss": 1.0135,
"step": 385
},
{
"epoch": 3.948849104859335,
"grad_norm": 0.08002277345563297,
"learning_rate": 1.9505729999918194e-05,
"loss": 1.0044,
"step": 386
},
{
"epoch": 3.959079283887468,
"grad_norm": 0.08738866259057297,
"learning_rate": 1.913010182747196e-05,
"loss": 1.0103,
"step": 387
},
{
"epoch": 3.969309462915601,
"grad_norm": 0.08459856390870199,
"learning_rate": 1.875763394885441e-05,
"loss": 1.0491,
"step": 388
},
{
"epoch": 3.979539641943734,
"grad_norm": 0.07894640689767103,
"learning_rate": 1.8388345702161556e-05,
"loss": 1.0355,
"step": 389
},
{
"epoch": 3.9897698209718673,
"grad_norm": 0.07643936790355793,
"learning_rate": 1.8022256260406756e-05,
"loss": 1.021,
"step": 390
},
{
"epoch": 4.0,
"grad_norm": 0.09661289153754662,
"learning_rate": 1.765938463052506e-05,
"loss": 1.288,
"step": 391
},
{
"epoch": 4.010230179028133,
"grad_norm": 0.08762181265193057,
"learning_rate": 1.729974965238651e-05,
"loss": 1.001,
"step": 392
},
{
"epoch": 4.020460358056266,
"grad_norm": 0.08647609139764888,
"learning_rate": 1.6943369997818066e-05,
"loss": 1.015,
"step": 393
},
{
"epoch": 4.030690537084399,
"grad_norm": 0.08788066772345464,
"learning_rate": 1.659026416963401e-05,
"loss": 1.0076,
"step": 394
},
{
"epoch": 4.040920716112532,
"grad_norm": 0.07750381562018518,
"learning_rate": 1.6240450500675393e-05,
"loss": 1.0148,
"step": 395
},
{
"epoch": 4.051150895140665,
"grad_norm": 0.07635694926329481,
"learning_rate": 1.5893947152858285e-05,
"loss": 1.0016,
"step": 396
},
{
"epoch": 4.061381074168798,
"grad_norm": 0.08148076505063037,
"learning_rate": 1.55507721162307e-05,
"loss": 1.0043,
"step": 397
},
{
"epoch": 4.071611253196931,
"grad_norm": 0.07751240865988411,
"learning_rate": 1.5210943208038634e-05,
"loss": 1.0288,
"step": 398
},
{
"epoch": 4.081841432225064,
"grad_norm": 0.07904341557445363,
"learning_rate": 1.4874478071801055e-05,
"loss": 1.0302,
"step": 399
},
{
"epoch": 4.092071611253197,
"grad_norm": 0.08036549522985713,
"learning_rate": 1.454139417639377e-05,
"loss": 1.0021,
"step": 400
},
{
"epoch": 4.10230179028133,
"grad_norm": 0.07597724481740194,
"learning_rate": 1.4211708815142599e-05,
"loss": 1.008,
"step": 401
},
{
"epoch": 4.112531969309463,
"grad_norm": 0.07493385254201147,
"learning_rate": 1.3885439104925387e-05,
"loss": 1.0082,
"step": 402
},
{
"epoch": 4.122762148337596,
"grad_norm": 0.08492760833944613,
"learning_rate": 1.3562601985283358e-05,
"loss": 1.0103,
"step": 403
},
{
"epoch": 4.132992327365729,
"grad_norm": 0.0819518128021678,
"learning_rate": 1.3243214217541751e-05,
"loss": 1.0186,
"step": 404
},
{
"epoch": 4.143222506393862,
"grad_norm": 0.07813515029694298,
"learning_rate": 1.2927292383939407e-05,
"loss": 1.0103,
"step": 405
},
{
"epoch": 4.153452685421995,
"grad_norm": 0.07571122126240162,
"learning_rate": 1.2614852886767932e-05,
"loss": 1.0172,
"step": 406
},
{
"epoch": 4.163682864450128,
"grad_norm": 0.07867401677851067,
"learning_rate": 1.2305911947520159e-05,
"loss": 1.0172,
"step": 407
},
{
"epoch": 4.173913043478261,
"grad_norm": 0.07427552294515202,
"learning_rate": 1.2000485606047837e-05,
"loss": 1.0254,
"step": 408
},
{
"epoch": 4.1841432225063935,
"grad_norm": 0.07918066825847282,
"learning_rate": 1.1698589719728911e-05,
"loss": 1.025,
"step": 409
},
{
"epoch": 4.194373401534527,
"grad_norm": 0.0745101976937812,
"learning_rate": 1.1400239962644294e-05,
"loss": 1.011,
"step": 410
},
{
"epoch": 4.20460358056266,
"grad_norm": 0.07635541094438233,
"learning_rate": 1.1105451824763933e-05,
"loss": 1.0064,
"step": 411
},
{
"epoch": 4.2148337595907925,
"grad_norm": 0.07409705314572482,
"learning_rate": 1.0814240611142765e-05,
"loss": 1.0049,
"step": 412
},
{
"epoch": 4.225063938618926,
"grad_norm": 0.07875185704588439,
"learning_rate": 1.0526621441125946e-05,
"loss": 1.0039,
"step": 413
},
{
"epoch": 4.235294117647059,
"grad_norm": 0.07576883769832815,
"learning_rate": 1.0242609247563924e-05,
"loss": 1.0204,
"step": 414
},
{
"epoch": 4.245524296675192,
"grad_norm": 0.07808157523983991,
"learning_rate": 9.962218776037234e-06,
"loss": 1.0178,
"step": 415
},
{
"epoch": 4.255754475703325,
"grad_norm": 0.0706208132598418,
"learning_rate": 9.68546458409077e-06,
"loss": 1.0144,
"step": 416
},
{
"epoch": 4.265984654731458,
"grad_norm": 0.0816397013567019,
"learning_rate": 9.41236104047806e-06,
"loss": 1.0008,
"step": 417
},
{
"epoch": 4.276214833759591,
"grad_norm": 0.07454955996130302,
"learning_rate": 9.14292232441528e-06,
"loss": 1.0102,
"step": 418
},
{
"epoch": 4.286445012787723,
"grad_norm": 0.0715010447940489,
"learning_rate": 8.877162424845012e-06,
"loss": 0.997,
"step": 419
},
{
"epoch": 4.296675191815857,
"grad_norm": 0.07752947453647589,
"learning_rate": 8.615095139710044e-06,
"loss": 1.0204,
"step": 420
},
{
"epoch": 4.30690537084399,
"grad_norm": 0.07999927748669688,
"learning_rate": 8.356734075236858e-06,
"loss": 1.0286,
"step": 421
},
{
"epoch": 4.3171355498721224,
"grad_norm": 0.07470822337555567,
"learning_rate": 8.102092645229392e-06,
"loss": 0.9999,
"step": 422
},
{
"epoch": 4.327365728900256,
"grad_norm": 0.06986020359409462,
"learning_rate": 7.8511840703725e-06,
"loss": 1.0157,
"step": 423
},
{
"epoch": 4.337595907928389,
"grad_norm": 0.07546615544304432,
"learning_rate": 7.604021377545518e-06,
"loss": 1.0177,
"step": 424
},
{
"epoch": 4.3478260869565215,
"grad_norm": 0.07303012390571562,
"learning_rate": 7.36061739914601e-06,
"loss": 1.0053,
"step": 425
},
{
"epoch": 4.358056265984655,
"grad_norm": 0.07983971604823199,
"learning_rate": 7.120984772423507e-06,
"loss": 1.0116,
"step": 426
},
{
"epoch": 4.368286445012788,
"grad_norm": 0.07412392301940252,
"learning_rate": 6.88513593882334e-06,
"loss": 1.0157,
"step": 427
},
{
"epoch": 4.378516624040921,
"grad_norm": 0.07540589914642663,
"learning_rate": 6.653083143340748e-06,
"loss": 1.0321,
"step": 428
},
{
"epoch": 4.388746803069053,
"grad_norm": 0.0734695901154156,
"learning_rate": 6.4248384338851146e-06,
"loss": 1.0166,
"step": 429
},
{
"epoch": 4.398976982097187,
"grad_norm": 0.07148917346240428,
"learning_rate": 6.2004136606544515e-06,
"loss": 1.0155,
"step": 430
},
{
"epoch": 4.40920716112532,
"grad_norm": 0.06576015271509353,
"learning_rate": 5.979820475520202e-06,
"loss": 1.0268,
"step": 431
},
{
"epoch": 4.419437340153452,
"grad_norm": 0.06602815168374586,
"learning_rate": 5.763070331422151e-06,
"loss": 1.0094,
"step": 432
},
{
"epoch": 4.429667519181586,
"grad_norm": 0.06806818912226788,
"learning_rate": 5.550174481773969e-06,
"loss": 1.0117,
"step": 433
},
{
"epoch": 4.439897698209719,
"grad_norm": 0.0654860198424932,
"learning_rate": 5.341143979878851e-06,
"loss": 1.024,
"step": 434
},
{
"epoch": 4.450127877237851,
"grad_norm": 0.06347915481210827,
"learning_rate": 5.135989678355664e-06,
"loss": 1.0068,
"step": 435
},
{
"epoch": 4.460358056265985,
"grad_norm": 0.06673007006524383,
"learning_rate": 4.934722228575481e-06,
"loss": 1.0144,
"step": 436
},
{
"epoch": 4.470588235294118,
"grad_norm": 0.06276490466658945,
"learning_rate": 4.7373520801085705e-06,
"loss": 1.0149,
"step": 437
},
{
"epoch": 4.4808184143222505,
"grad_norm": 0.06462800206504453,
"learning_rate": 4.543889480181944e-06,
"loss": 1.0209,
"step": 438
},
{
"epoch": 4.491048593350383,
"grad_norm": 0.06638702716213918,
"learning_rate": 4.354344473147194e-06,
"loss": 1.0229,
"step": 439
},
{
"epoch": 4.501278772378517,
"grad_norm": 0.06786708229119566,
"learning_rate": 4.1687268999591164e-06,
"loss": 1.0093,
"step": 440
},
{
"epoch": 4.5115089514066495,
"grad_norm": 0.06344929556504937,
"learning_rate": 3.98704639766474e-06,
"loss": 1.0227,
"step": 441
},
{
"epoch": 4.521739130434782,
"grad_norm": 0.061257597771025976,
"learning_rate": 3.809312398903e-06,
"loss": 1.0206,
"step": 442
},
{
"epoch": 4.531969309462916,
"grad_norm": 0.06207018671567817,
"learning_rate": 3.6355341314149216e-06,
"loss": 1.0061,
"step": 443
},
{
"epoch": 4.542199488491049,
"grad_norm": 0.0697177559048247,
"learning_rate": 3.465720617564676e-06,
"loss": 1.0001,
"step": 444
},
{
"epoch": 4.552429667519181,
"grad_norm": 0.06909570953584229,
"learning_rate": 3.299880673871023e-06,
"loss": 1.0179,
"step": 445
},
{
"epoch": 4.562659846547315,
"grad_norm": 0.0660073639280078,
"learning_rate": 3.138022910549632e-06,
"loss": 1.0261,
"step": 446
},
{
"epoch": 4.572890025575448,
"grad_norm": 0.061208022344978574,
"learning_rate": 2.980155731066017e-06,
"loss": 0.9983,
"step": 447
},
{
"epoch": 4.58312020460358,
"grad_norm": 0.05925952496152496,
"learning_rate": 2.8262873316992556e-06,
"loss": 1.0232,
"step": 448
},
{
"epoch": 4.593350383631714,
"grad_norm": 0.06402981619720209,
"learning_rate": 2.676425701116463e-06,
"loss": 1.0065,
"step": 449
},
{
"epoch": 4.603580562659847,
"grad_norm": 0.058782758741933706,
"learning_rate": 2.530578619957993e-06,
"loss": 1.0117,
"step": 450
},
{
"epoch": 4.6138107416879794,
"grad_norm": 0.05836448954212346,
"learning_rate": 2.3887536604334784e-06,
"loss": 0.9904,
"step": 451
},
{
"epoch": 4.624040920716112,
"grad_norm": 0.05866978691158495,
"learning_rate": 2.2509581859287576e-06,
"loss": 1.018,
"step": 452
},
{
"epoch": 4.634271099744246,
"grad_norm": 0.05897537411295091,
"learning_rate": 2.117199350623462e-06,
"loss": 1.0224,
"step": 453
},
{
"epoch": 4.6445012787723785,
"grad_norm": 0.0576826219117585,
"learning_rate": 1.987484099119712e-06,
"loss": 1.0256,
"step": 454
},
{
"epoch": 4.654731457800511,
"grad_norm": 0.058074308794142736,
"learning_rate": 1.8618191660814356e-06,
"loss": 1.0126,
"step": 455
},
{
"epoch": 4.664961636828645,
"grad_norm": 0.057718766623449665,
"learning_rate": 1.7402110758847834e-06,
"loss": 1.0064,
"step": 456
},
{
"epoch": 4.675191815856778,
"grad_norm": 0.055268606367167225,
"learning_rate": 1.6226661422794033e-06,
"loss": 1.0015,
"step": 457
},
{
"epoch": 4.68542199488491,
"grad_norm": 0.06120460420865381,
"learning_rate": 1.5091904680605862e-06,
"loss": 1.0195,
"step": 458
},
{
"epoch": 4.695652173913043,
"grad_norm": 0.057198583474872174,
"learning_rate": 1.3997899447524277e-06,
"loss": 1.0313,
"step": 459
},
{
"epoch": 4.705882352941177,
"grad_norm": 0.054125320926830395,
"learning_rate": 1.294470252302009e-06,
"loss": 1.0074,
"step": 460
},
{
"epoch": 4.716112531969309,
"grad_norm": 0.05767289138418469,
"learning_rate": 1.193236858784408e-06,
"loss": 1.0073,
"step": 461
},
{
"epoch": 4.726342710997442,
"grad_norm": 0.058374045513090105,
"learning_rate": 1.0960950201188524e-06,
"loss": 1.0217,
"step": 462
},
{
"epoch": 4.736572890025576,
"grad_norm": 0.05554209835620826,
"learning_rate": 1.003049779795866e-06,
"loss": 1.0167,
"step": 463
},
{
"epoch": 4.746803069053708,
"grad_norm": 0.06184311865928885,
"learning_rate": 9.141059686153419e-07,
"loss": 1.0207,
"step": 464
},
{
"epoch": 4.757033248081841,
"grad_norm": 0.056079764830628674,
"learning_rate": 8.292682044358114e-07,
"loss": 1.0169,
"step": 465
},
{
"epoch": 4.767263427109975,
"grad_norm": 0.05903827365785868,
"learning_rate": 7.485408919346171e-07,
"loss": 1.0276,
"step": 466
},
{
"epoch": 4.7774936061381075,
"grad_norm": 0.05665451559441624,
"learning_rate": 6.719282223793056e-07,
"loss": 1.0108,
"step": 467
},
{
"epoch": 4.78772378516624,
"grad_norm": 0.05541913708216884,
"learning_rate": 5.994341734099429e-07,
"loss": 1.0213,
"step": 468
},
{
"epoch": 4.797953964194374,
"grad_norm": 0.0531727158843823,
"learning_rate": 5.310625088326671e-07,
"loss": 0.9962,
"step": 469
},
{
"epoch": 4.8081841432225065,
"grad_norm": 0.056474278323680495,
"learning_rate": 4.6681677842421724e-07,
"loss": 1.0079,
"step": 470
},
{
"epoch": 4.818414322250639,
"grad_norm": 0.05529573865410167,
"learning_rate": 4.067003177476991e-07,
"loss": 1.0025,
"step": 471
},
{
"epoch": 4.828644501278772,
"grad_norm": 0.055929732769464856,
"learning_rate": 3.507162479793369e-07,
"loss": 1.0173,
"step": 472
},
{
"epoch": 4.838874680306906,
"grad_norm": 0.0546056253906124,
"learning_rate": 2.9886747574646936e-07,
"loss": 1.0001,
"step": 473
},
{
"epoch": 4.849104859335038,
"grad_norm": 0.05682769159587846,
"learning_rate": 2.511566929766396e-07,
"loss": 1.0062,
"step": 474
},
{
"epoch": 4.859335038363171,
"grad_norm": 0.05365397819606409,
"learning_rate": 2.075863767577957e-07,
"loss": 1.0195,
"step": 475
},
{
"epoch": 4.869565217391305,
"grad_norm": 0.05538514973032816,
"learning_rate": 1.681587892097536e-07,
"loss": 1.0159,
"step": 476
},
{
"epoch": 4.879795396419437,
"grad_norm": 0.05391134086338163,
"learning_rate": 1.3287597736667323e-07,
"loss": 1.0233,
"step": 477
},
{
"epoch": 4.89002557544757,
"grad_norm": 0.054055617154126184,
"learning_rate": 1.0173977307082361e-07,
"loss": 1.0188,
"step": 478
},
{
"epoch": 4.900255754475703,
"grad_norm": 0.05372397484026782,
"learning_rate": 7.475179287748547e-08,
"loss": 1.0235,
"step": 479
},
{
"epoch": 4.910485933503836,
"grad_norm": 0.054794275718811285,
"learning_rate": 5.191343797096515e-08,
"loss": 1.0018,
"step": 480
},
{
"epoch": 4.920716112531969,
"grad_norm": 0.05350058384849903,
"learning_rate": 3.322589409190613e-08,
"loss": 1.009,
"step": 481
},
{
"epoch": 4.930946291560103,
"grad_norm": 0.05480646007195536,
"learning_rate": 1.8690131475711527e-08,
"loss": 1.0273,
"step": 482
},
{
"epoch": 4.9411764705882355,
"grad_norm": 0.05416557096319236,
"learning_rate": 8.306904802148907e-09,
"loss": 1.0322,
"step": 483
},
{
"epoch": 4.951406649616368,
"grad_norm": 0.053492193014158126,
"learning_rate": 2.07675315618161e-09,
"loss": 1.0035,
"step": 484
},
{
"epoch": 4.961636828644501,
"grad_norm": 0.055105663696246775,
"learning_rate": 0.0,
"loss": 1.0124,
"step": 485
},
{
"epoch": 4.961636828644501,
"step": 485,
"total_flos": 7065760181780480.0,
"train_loss": 1.075610858509221,
"train_runtime": 69151.8205,
"train_samples_per_second": 7.23,
"train_steps_per_second": 0.007
}
],
"logging_steps": 1.0,
"max_steps": 485,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 7065760181780480.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}