Files
XiYanSQL-QwenCoder-3B-2502/trainer_state.json
ModelHub XC 40c517c0cf 初始化项目,由ModelHub XC社区提供模型
Model: XGenerationLab/XiYanSQL-QwenCoder-3B-2502
Source: Original Platform
2026-05-07 19:10:10 +08:00

28090 lines
636 KiB
JSON

{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.5296367112810707,
"eval_steps": 500,
"global_step": 4000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0003824091778202677,
"grad_norm": 8.127729221609483,
"learning_rate": 0.0,
"loss": 0.7499,
"step": 1
},
{
"epoch": 0.0007648183556405354,
"grad_norm": 7.963595529643983,
"learning_rate": 9.658899046395883e-08,
"loss": 0.7745,
"step": 2
},
{
"epoch": 0.001147227533460803,
"grad_norm": 6.399453926631711,
"learning_rate": 1.530899278678881e-07,
"loss": 0.827,
"step": 3
},
{
"epoch": 0.0015296367112810707,
"grad_norm": 7.212906755583777,
"learning_rate": 1.9317798092791765e-07,
"loss": 0.9354,
"step": 4
},
{
"epoch": 0.0019120458891013384,
"grad_norm": 4.403583991640848,
"learning_rate": 2.2427269061507353e-07,
"loss": 0.8108,
"step": 5
},
{
"epoch": 0.002294455066921606,
"grad_norm": 9.30785115586022,
"learning_rate": 2.496789183318469e-07,
"loss": 0.8121,
"step": 6
},
{
"epoch": 0.0026768642447418736,
"grad_norm": 10.972546390514474,
"learning_rate": 2.711595777955698e-07,
"loss": 0.8276,
"step": 7
},
{
"epoch": 0.0030592734225621415,
"grad_norm": 7.964528600005935,
"learning_rate": 2.8976697139187646e-07,
"loss": 0.889,
"step": 8
},
{
"epoch": 0.0034416826003824093,
"grad_norm": 11.538619501059664,
"learning_rate": 3.061798557357762e-07,
"loss": 0.5456,
"step": 9
},
{
"epoch": 0.0038240917782026767,
"grad_norm": 9.923505802246833,
"learning_rate": 3.208616810790324e-07,
"loss": 0.5617,
"step": 10
},
{
"epoch": 0.004206500956022944,
"grad_norm": 13.868230594571852,
"learning_rate": 3.341430076232756e-07,
"loss": 0.8914,
"step": 11
},
{
"epoch": 0.004588910133843212,
"grad_norm": 15.82952470982823,
"learning_rate": 3.4626790879580574e-07,
"loss": 0.3943,
"step": 12
},
{
"epoch": 0.00497131931166348,
"grad_norm": 13.881207151660286,
"learning_rate": 3.5742173664798447e-07,
"loss": 0.9886,
"step": 13
},
{
"epoch": 0.005353728489483747,
"grad_norm": 9.351576777218634,
"learning_rate": 3.677485682595286e-07,
"loss": 1.2225,
"step": 14
},
{
"epoch": 0.0057361376673040155,
"grad_norm": 6.921096792456155,
"learning_rate": 3.773626184829616e-07,
"loss": 0.6324,
"step": 15
},
{
"epoch": 0.006118546845124283,
"grad_norm": 7.911741257096208,
"learning_rate": 3.863559618558353e-07,
"loss": 1.0263,
"step": 16
},
{
"epoch": 0.00650095602294455,
"grad_norm": 9.023578698841147,
"learning_rate": 3.948039093953151e-07,
"loss": 0.8379,
"step": 17
},
{
"epoch": 0.006883365200764819,
"grad_norm": 5.773447080611977,
"learning_rate": 4.02768846199735e-07,
"loss": 0.6313,
"step": 18
},
{
"epoch": 0.007265774378585086,
"grad_norm": 6.391540312564622,
"learning_rate": 4.103030300875908e-07,
"loss": 0.6562,
"step": 19
},
{
"epoch": 0.0076481835564053535,
"grad_norm": 4.734698092574861,
"learning_rate": 4.174506715429912e-07,
"loss": 0.9944,
"step": 20
},
{
"epoch": 0.00803059273422562,
"grad_norm": 5.543325744703233,
"learning_rate": 4.242495056634579e-07,
"loss": 0.8748,
"step": 21
},
{
"epoch": 0.008413001912045888,
"grad_norm": 6.139313458259769,
"learning_rate": 4.3073199808723443e-07,
"loss": 0.6497,
"step": 22
},
{
"epoch": 0.008795411089866157,
"grad_norm": 10.51882881230223,
"learning_rate": 4.369262826367178e-07,
"loss": 0.7235,
"step": 23
},
{
"epoch": 0.009177820267686425,
"grad_norm": 9.14158481873066,
"learning_rate": 4.428568992597646e-07,
"loss": 0.4482,
"step": 24
},
{
"epoch": 0.009560229445506692,
"grad_norm": 10.990275506131937,
"learning_rate": 4.4854538123014706e-07,
"loss": 0.8668,
"step": 25
},
{
"epoch": 0.00994263862332696,
"grad_norm": 11.815499706960994,
"learning_rate": 4.540107271119433e-07,
"loss": 0.8262,
"step": 26
},
{
"epoch": 0.010325047801147227,
"grad_norm": 7.542823857279405,
"learning_rate": 4.592697836036643e-07,
"loss": 0.9995,
"step": 27
},
{
"epoch": 0.010707456978967494,
"grad_norm": 3.439453125,
"learning_rate": 4.6433755872348744e-07,
"loss": 0.652,
"step": 28
},
{
"epoch": 0.011089866156787764,
"grad_norm": 4.447922408065247,
"learning_rate": 4.6922748001247033e-07,
"loss": 0.4116,
"step": 29
},
{
"epoch": 0.011472275334608031,
"grad_norm": 3.5419146432268502,
"learning_rate": 4.739516089469205e-07,
"loss": 0.603,
"step": 30
},
{
"epoch": 0.011854684512428298,
"grad_norm": 5.595453269399992,
"learning_rate": 4.78520820180538e-07,
"loss": 0.8754,
"step": 31
},
{
"epoch": 0.012237093690248566,
"grad_norm": 4.664053298711087,
"learning_rate": 4.829449523197942e-07,
"loss": 0.728,
"step": 32
},
{
"epoch": 0.012619502868068833,
"grad_norm": 4.184285994388638,
"learning_rate": 4.872329354911637e-07,
"loss": 0.7668,
"step": 33
},
{
"epoch": 0.0130019120458891,
"grad_norm": 3.8462487759245723,
"learning_rate": 4.91392899859274e-07,
"loss": 0.7642,
"step": 34
},
{
"epoch": 0.01338432122370937,
"grad_norm": 5.035267807203385,
"learning_rate": 4.954322684106434e-07,
"loss": 0.4231,
"step": 35
},
{
"epoch": 0.013766730401529637,
"grad_norm": 4.700427932737028,
"learning_rate": 4.993578366636938e-07,
"loss": 0.4554,
"step": 36
},
{
"epoch": 0.014149139579349905,
"grad_norm": 6.65272564780195,
"learning_rate": 5.031758414551729e-07,
"loss": 0.3467,
"step": 37
},
{
"epoch": 0.014531548757170172,
"grad_norm": 6.540097900964421,
"learning_rate": 5.068920205515497e-07,
"loss": 0.5521,
"step": 38
},
{
"epoch": 0.01491395793499044,
"grad_norm": 6.187611588521872,
"learning_rate": 5.105116645158726e-07,
"loss": 0.8477,
"step": 39
},
{
"epoch": 0.015296367112810707,
"grad_norm": 4.212933408814397,
"learning_rate": 5.140396620069501e-07,
"loss": 0.8013,
"step": 40
},
{
"epoch": 0.015678776290630976,
"grad_norm": 3.7306584334956354,
"learning_rate": 5.174805394842197e-07,
"loss": 0.34,
"step": 41
},
{
"epoch": 0.01606118546845124,
"grad_norm": 3.8129084634065613,
"learning_rate": 5.208384961274167e-07,
"loss": 0.7891,
"step": 42
},
{
"epoch": 0.01644359464627151,
"grad_norm": 3.521126067201788,
"learning_rate": 5.241174346468368e-07,
"loss": 0.8447,
"step": 43
},
{
"epoch": 0.016826003824091777,
"grad_norm": 5.279756052566196,
"learning_rate": 5.273209885511933e-07,
"loss": 0.6151,
"step": 44
},
{
"epoch": 0.017208413001912046,
"grad_norm": 4.009723289626729,
"learning_rate": 5.304525463508497e-07,
"loss": 0.6095,
"step": 45
},
{
"epoch": 0.017590822179732315,
"grad_norm": 3.7728758842220267,
"learning_rate": 5.335152731006765e-07,
"loss": 0.3871,
"step": 46
},
{
"epoch": 0.01797323135755258,
"grad_norm": 4.866870067648313,
"learning_rate": 5.365121296259033e-07,
"loss": 0.4796,
"step": 47
},
{
"epoch": 0.01835564053537285,
"grad_norm": 4.8744050665499286,
"learning_rate": 5.394458897237234e-07,
"loss": 0.5812,
"step": 48
},
{
"epoch": 0.018738049713193115,
"grad_norm": 9.108527378017467,
"learning_rate": 5.423191555911396e-07,
"loss": 0.319,
"step": 49
},
{
"epoch": 0.019120458891013385,
"grad_norm": 9.158412014105785,
"learning_rate": 5.45134371694106e-07,
"loss": 0.7643,
"step": 50
},
{
"epoch": 0.019502868068833654,
"grad_norm": 5.790040627239837,
"learning_rate": 5.478938372632032e-07,
"loss": 0.6109,
"step": 51
},
{
"epoch": 0.01988527724665392,
"grad_norm": 6.096683676341199,
"learning_rate": 5.505997175759021e-07,
"loss": 0.9369,
"step": 52
},
{
"epoch": 0.02026768642447419,
"grad_norm": 4.441546891668887,
"learning_rate": 5.532540541641197e-07,
"loss": 0.7858,
"step": 53
},
{
"epoch": 0.020650095602294454,
"grad_norm": 2.8696014470488382,
"learning_rate": 5.558587740676232e-07,
"loss": 0.8667,
"step": 54
},
{
"epoch": 0.021032504780114723,
"grad_norm": 3.2738861286399046,
"learning_rate": 5.584156982383491e-07,
"loss": 0.4673,
"step": 55
},
{
"epoch": 0.02141491395793499,
"grad_norm": 3.697994492835673,
"learning_rate": 5.609265491874463e-07,
"loss": 0.6328,
"step": 56
},
{
"epoch": 0.021797323135755258,
"grad_norm": 3.125139767382225,
"learning_rate": 5.633929579554789e-07,
"loss": 0.4692,
"step": 57
},
{
"epoch": 0.022179732313575527,
"grad_norm": 4.298443428237915,
"learning_rate": 5.658164704764291e-07,
"loss": 0.6768,
"step": 58
},
{
"epoch": 0.022562141491395793,
"grad_norm": 4.398677297665077,
"learning_rate": 5.681985533976847e-07,
"loss": 0.6619,
"step": 59
},
{
"epoch": 0.022944550669216062,
"grad_norm": 5.818330527160778,
"learning_rate": 5.705405994108792e-07,
"loss": 0.6268,
"step": 60
},
{
"epoch": 0.023326959847036328,
"grad_norm": 5.36324426425703,
"learning_rate": 5.728439321421063e-07,
"loss": 0.3157,
"step": 61
},
{
"epoch": 0.023709369024856597,
"grad_norm": 5.248781153565317,
"learning_rate": 5.751098106444967e-07,
"loss": 0.3525,
"step": 62
},
{
"epoch": 0.024091778202676863,
"grad_norm": 6.7358008908634925,
"learning_rate": 5.77339433531346e-07,
"loss": 0.6165,
"step": 63
},
{
"epoch": 0.024474187380497132,
"grad_norm": 5.135947096269937,
"learning_rate": 5.795339427837529e-07,
"loss": 0.9469,
"step": 64
},
{
"epoch": 0.0248565965583174,
"grad_norm": 4.721443670581346,
"learning_rate": 5.816944272630579e-07,
"loss": 1.0586,
"step": 65
},
{
"epoch": 0.025239005736137667,
"grad_norm": 4.560383135778482,
"learning_rate": 5.838219259551224e-07,
"loss": 0.6226,
"step": 66
},
{
"epoch": 0.025621414913957936,
"grad_norm": 3.0533395900689975,
"learning_rate": 5.859174309706495e-07,
"loss": 0.5508,
"step": 67
},
{
"epoch": 0.0260038240917782,
"grad_norm": 4.38574534050627,
"learning_rate": 5.879818903232327e-07,
"loss": 0.6519,
"step": 68
},
{
"epoch": 0.02638623326959847,
"grad_norm": 3.1760295850591347,
"learning_rate": 5.900162105046059e-07,
"loss": 0.7339,
"step": 69
},
{
"epoch": 0.02676864244741874,
"grad_norm": 3.243632755108415,
"learning_rate": 5.920212588746022e-07,
"loss": 0.4918,
"step": 70
},
{
"epoch": 0.027151051625239005,
"grad_norm": 3.432834024756851,
"learning_rate": 5.939978658815961e-07,
"loss": 0.5308,
"step": 71
},
{
"epoch": 0.027533460803059275,
"grad_norm": 3.677125723417059,
"learning_rate": 5.959468271276526e-07,
"loss": 0.3372,
"step": 72
},
{
"epoch": 0.02791586998087954,
"grad_norm": 2.8933168816281856,
"learning_rate": 5.978689052912401e-07,
"loss": 0.3834,
"step": 73
},
{
"epoch": 0.02829827915869981,
"grad_norm": 4.377616318007501,
"learning_rate": 5.997648319191317e-07,
"loss": 0.285,
"step": 74
},
{
"epoch": 0.028680688336520075,
"grad_norm": 5.329271657395167,
"learning_rate": 6.016353090980351e-07,
"loss": 0.4062,
"step": 75
},
{
"epoch": 0.029063097514340344,
"grad_norm": 7.870238302614566,
"learning_rate": 6.034810110155085e-07,
"loss": 0.7765,
"step": 76
},
{
"epoch": 0.029445506692160613,
"grad_norm": 4.2678637096949625,
"learning_rate": 6.053025854188454e-07,
"loss": 0.7735,
"step": 77
},
{
"epoch": 0.02982791586998088,
"grad_norm": 3.691418909121812,
"learning_rate": 6.071006549798314e-07,
"loss": 0.7857,
"step": 78
},
{
"epoch": 0.030210325047801148,
"grad_norm": 2.8591930560816983,
"learning_rate": 6.088758185725654e-07,
"loss": 0.7999,
"step": 79
},
{
"epoch": 0.030592734225621414,
"grad_norm": 3.9392785793477474,
"learning_rate": 6.106286524709088e-07,
"loss": 0.7652,
"step": 80
},
{
"epoch": 0.030975143403441683,
"grad_norm": 4.318070062532729,
"learning_rate": 6.123597114715524e-07,
"loss": 0.6564,
"step": 81
},
{
"epoch": 0.03135755258126195,
"grad_norm": 3.853141414610464,
"learning_rate": 6.140695299481785e-07,
"loss": 0.6589,
"step": 82
},
{
"epoch": 0.03173996175908222,
"grad_norm": 3.454451500377018,
"learning_rate": 6.157586228417297e-07,
"loss": 0.6253,
"step": 83
},
{
"epoch": 0.03212237093690248,
"grad_norm": 3.244831082924636,
"learning_rate": 6.174274865913755e-07,
"loss": 0.2258,
"step": 84
},
{
"epoch": 0.032504780114722756,
"grad_norm": 4.9841997360937835,
"learning_rate": 6.190766000103886e-07,
"loss": 0.4724,
"step": 85
},
{
"epoch": 0.03288718929254302,
"grad_norm": 4.358109922986119,
"learning_rate": 6.207064251107957e-07,
"loss": 0.3161,
"step": 86
},
{
"epoch": 0.03326959847036329,
"grad_norm": 3.752613174388839,
"learning_rate": 6.223174078803584e-07,
"loss": 0.2185,
"step": 87
},
{
"epoch": 0.03365200764818355,
"grad_norm": 6.035992748966501,
"learning_rate": 6.239099790151521e-07,
"loss": 0.7716,
"step": 88
},
{
"epoch": 0.034034416826003826,
"grad_norm": 5.726135487788577,
"learning_rate": 6.254845546107528e-07,
"loss": 1.1187,
"step": 89
},
{
"epoch": 0.03441682600382409,
"grad_norm": 4.188566157565,
"learning_rate": 6.270415368148085e-07,
"loss": 0.6178,
"step": 90
},
{
"epoch": 0.03479923518164436,
"grad_norm": 1.8973622357764504,
"learning_rate": 6.285813144435542e-07,
"loss": 0.82,
"step": 91
},
{
"epoch": 0.03518164435946463,
"grad_norm": 3.203917465804253,
"learning_rate": 6.301042635646355e-07,
"loss": 0.7247,
"step": 92
},
{
"epoch": 0.035564053537284895,
"grad_norm": 4.069173173828255,
"learning_rate": 6.316107480484261e-07,
"loss": 0.9489,
"step": 93
},
{
"epoch": 0.03594646271510516,
"grad_norm": 5.255571224914232,
"learning_rate": 6.331011200898622e-07,
"loss": 0.3546,
"step": 94
},
{
"epoch": 0.036328871892925434,
"grad_norm": 2.9958642426559994,
"learning_rate": 6.345757207026644e-07,
"loss": 0.5467,
"step": 95
},
{
"epoch": 0.0367112810707457,
"grad_norm": 3.6217326204556106,
"learning_rate": 6.360348801876822e-07,
"loss": 0.2342,
"step": 96
},
{
"epoch": 0.037093690248565965,
"grad_norm": 4.380818231179398,
"learning_rate": 6.374789185769719e-07,
"loss": 0.2475,
"step": 97
},
{
"epoch": 0.03747609942638623,
"grad_norm": 3.792650297804044,
"learning_rate": 6.389081460550985e-07,
"loss": 0.3622,
"step": 98
},
{
"epoch": 0.0378585086042065,
"grad_norm": 4.318067191395842,
"learning_rate": 6.403228633590517e-07,
"loss": 0.2876,
"step": 99
},
{
"epoch": 0.03824091778202677,
"grad_norm": 4.3852408306664685,
"learning_rate": 6.417233621580648e-07,
"loss": 0.3416,
"step": 100
},
{
"epoch": 0.038623326959847035,
"grad_norm": 5.55577400837663,
"learning_rate": 6.431099254145344e-07,
"loss": 0.6162,
"step": 101
},
{
"epoch": 0.03900573613766731,
"grad_norm": 4.679812829276982,
"learning_rate": 6.44482827727162e-07,
"loss": 0.8162,
"step": 102
},
{
"epoch": 0.03938814531548757,
"grad_norm": 4.147621318751239,
"learning_rate": 6.458423356573556e-07,
"loss": 1.1291,
"step": 103
},
{
"epoch": 0.03977055449330784,
"grad_norm": 3.2625495994568445,
"learning_rate": 6.471887080398609e-07,
"loss": 0.6366,
"step": 104
},
{
"epoch": 0.040152963671128104,
"grad_norm": 3.2876321396179984,
"learning_rate": 6.485221962785314e-07,
"loss": 0.877,
"step": 105
},
{
"epoch": 0.04053537284894838,
"grad_norm": 3.51584336344417,
"learning_rate": 6.498430446280784e-07,
"loss": 0.4054,
"step": 106
},
{
"epoch": 0.04091778202676864,
"grad_norm": 3.4273604229777384,
"learning_rate": 6.511514904625936e-07,
"loss": 0.4965,
"step": 107
},
{
"epoch": 0.04130019120458891,
"grad_norm": 3.827336942561945,
"learning_rate": 6.52447764531582e-07,
"loss": 0.5168,
"step": 108
},
{
"epoch": 0.04168260038240918,
"grad_norm": 3.217148613937907,
"learning_rate": 6.537320912041942e-07,
"loss": 0.5542,
"step": 109
},
{
"epoch": 0.04206500956022945,
"grad_norm": 5.320029812779465,
"learning_rate": 6.55004688702308e-07,
"loss": 0.4252,
"step": 110
},
{
"epoch": 0.04244741873804971,
"grad_norm": 3.4351211639757118,
"learning_rate": 6.562657693230609e-07,
"loss": 0.3598,
"step": 111
},
{
"epoch": 0.04282982791586998,
"grad_norm": 4.734678353137203,
"learning_rate": 6.575155396514051e-07,
"loss": 0.3292,
"step": 112
},
{
"epoch": 0.04321223709369025,
"grad_norm": 6.986725618584486,
"learning_rate": 6.587542007632132e-07,
"loss": 0.8562,
"step": 113
},
{
"epoch": 0.043594646271510516,
"grad_norm": 4.767342479966072,
"learning_rate": 6.599819484194378e-07,
"loss": 0.9069,
"step": 114
},
{
"epoch": 0.04397705544933078,
"grad_norm": 5.708434092428296,
"learning_rate": 6.611989732517913e-07,
"loss": 0.7799,
"step": 115
},
{
"epoch": 0.044359464627151055,
"grad_norm": 2.143719184142411,
"learning_rate": 6.62405460940388e-07,
"loss": 0.762,
"step": 116
},
{
"epoch": 0.04474187380497132,
"grad_norm": 3.075753164708183,
"learning_rate": 6.636015923837607e-07,
"loss": 0.6046,
"step": 117
},
{
"epoch": 0.045124282982791586,
"grad_norm": 4.041382586271456,
"learning_rate": 6.647875438616435e-07,
"loss": 0.9243,
"step": 118
},
{
"epoch": 0.04550669216061185,
"grad_norm": 4.0781175167555865,
"learning_rate": 6.659634871908849e-07,
"loss": 0.6291,
"step": 119
},
{
"epoch": 0.045889101338432124,
"grad_norm": 2.947617658316482,
"learning_rate": 6.671295898748382e-07,
"loss": 0.6508,
"step": 120
},
{
"epoch": 0.04627151051625239,
"grad_norm": 3.7795709081852134,
"learning_rate": 6.682860152465512e-07,
"loss": 0.4124,
"step": 121
},
{
"epoch": 0.046653919694072656,
"grad_norm": 3.5289342257995586,
"learning_rate": 6.694329226060651e-07,
"loss": 0.4992,
"step": 122
},
{
"epoch": 0.04703632887189293,
"grad_norm": 4.706878684962685,
"learning_rate": 6.705704673521077e-07,
"loss": 0.2933,
"step": 123
},
{
"epoch": 0.047418738049713194,
"grad_norm": 5.110256294049169,
"learning_rate": 6.716988011084555e-07,
"loss": 0.4063,
"step": 124
},
{
"epoch": 0.04780114722753346,
"grad_norm": 4.520100734904757,
"learning_rate": 6.728180718452207e-07,
"loss": 0.4,
"step": 125
},
{
"epoch": 0.048183556405353725,
"grad_norm": 7.432639962375598,
"learning_rate": 6.739284239953048e-07,
"loss": 0.695,
"step": 126
},
{
"epoch": 0.048565965583174,
"grad_norm": 4.290670013747284,
"learning_rate": 6.750299985662519e-07,
"loss": 0.4686,
"step": 127
},
{
"epoch": 0.048948374760994263,
"grad_norm": 4.121158313697451,
"learning_rate": 6.761229332477118e-07,
"loss": 0.6365,
"step": 128
},
{
"epoch": 0.04933078393881453,
"grad_norm": 2.135314479957041,
"learning_rate": 6.772073625147248e-07,
"loss": 0.6846,
"step": 129
},
{
"epoch": 0.0497131931166348,
"grad_norm": 2.342096991950194,
"learning_rate": 6.782834177270168e-07,
"loss": 0.6111,
"step": 130
},
{
"epoch": 0.05009560229445507,
"grad_norm": 3.647894547310256,
"learning_rate": 6.793512272244895e-07,
"loss": 0.4072,
"step": 131
},
{
"epoch": 0.05047801147227533,
"grad_norm": 4.354125440971228,
"learning_rate": 6.804109164190813e-07,
"loss": 0.6969,
"step": 132
},
{
"epoch": 0.050860420650095606,
"grad_norm": 2.7346670812193903,
"learning_rate": 6.814626078831606e-07,
"loss": 0.5829,
"step": 133
},
{
"epoch": 0.05124282982791587,
"grad_norm": 4.094534063602115,
"learning_rate": 6.825064214346084e-07,
"loss": 0.3749,
"step": 134
},
{
"epoch": 0.05162523900573614,
"grad_norm": 3.382758089469319,
"learning_rate": 6.835424742187379e-07,
"loss": 0.3332,
"step": 135
},
{
"epoch": 0.0520076481835564,
"grad_norm": 2.683021984242027,
"learning_rate": 6.845708807871916e-07,
"loss": 0.1764,
"step": 136
},
{
"epoch": 0.052390057361376675,
"grad_norm": 4.509963555368485,
"learning_rate": 6.855917531739482e-07,
"loss": 0.2143,
"step": 137
},
{
"epoch": 0.05277246653919694,
"grad_norm": 5.256642816586635,
"learning_rate": 6.866052009685647e-07,
"loss": 0.4351,
"step": 138
},
{
"epoch": 0.05315487571701721,
"grad_norm": 5.834538834758634,
"learning_rate": 6.876113313867756e-07,
"loss": 0.6942,
"step": 139
},
{
"epoch": 0.05353728489483748,
"grad_norm": 3.014295690206189,
"learning_rate": 6.88610249338561e-07,
"loss": 0.4463,
"step": 140
},
{
"epoch": 0.053919694072657745,
"grad_norm": 2.2857408607164182,
"learning_rate": 6.896020574937915e-07,
"loss": 0.5019,
"step": 141
},
{
"epoch": 0.05430210325047801,
"grad_norm": 2.8317586412436886,
"learning_rate": 6.905868563455549e-07,
"loss": 0.551,
"step": 142
},
{
"epoch": 0.054684512428298276,
"grad_norm": 3.590807962694632,
"learning_rate": 6.9156474427126e-07,
"loss": 0.5001,
"step": 143
},
{
"epoch": 0.05506692160611855,
"grad_norm": 3.3447864387888737,
"learning_rate": 6.925358175916115e-07,
"loss": 0.6405,
"step": 144
},
{
"epoch": 0.055449330783938815,
"grad_norm": 3.5812648007285097,
"learning_rate": 6.935001706275438e-07,
"loss": 0.3969,
"step": 145
},
{
"epoch": 0.05583173996175908,
"grad_norm": 2.588371307068287,
"learning_rate": 6.944578957551989e-07,
"loss": 0.4425,
"step": 146
},
{
"epoch": 0.05621414913957935,
"grad_norm": 3.8032063258431816,
"learning_rate": 6.954090834590278e-07,
"loss": 0.5201,
"step": 147
},
{
"epoch": 0.05659655831739962,
"grad_norm": 3.972727066954228,
"learning_rate": 6.963538223830905e-07,
"loss": 0.4198,
"step": 148
},
{
"epoch": 0.056978967495219884,
"grad_norm": 4.727081825163079,
"learning_rate": 6.972921993806316e-07,
"loss": 0.3914,
"step": 149
},
{
"epoch": 0.05736137667304015,
"grad_norm": 3.8566009806911032,
"learning_rate": 6.982242995619941e-07,
"loss": 0.3418,
"step": 150
},
{
"epoch": 0.05774378585086042,
"grad_norm": 9.883895151296194,
"learning_rate": 6.991502063409444e-07,
"loss": 0.4184,
"step": 151
},
{
"epoch": 0.05812619502868069,
"grad_norm": 3.9589817252241994,
"learning_rate": 7.000700014794673e-07,
"loss": 0.9268,
"step": 152
},
{
"epoch": 0.058508604206500954,
"grad_norm": 3.6920673081202664,
"learning_rate": 7.009837651310913e-07,
"loss": 0.7311,
"step": 153
},
{
"epoch": 0.05889101338432123,
"grad_norm": 2.8818112834107232,
"learning_rate": 7.018915758828043e-07,
"loss": 0.507,
"step": 154
},
{
"epoch": 0.05927342256214149,
"grad_norm": 4.095078580332929,
"learning_rate": 7.027935107956114e-07,
"loss": 0.6913,
"step": 155
},
{
"epoch": 0.05965583173996176,
"grad_norm": 4.097583635557744,
"learning_rate": 7.036896454437902e-07,
"loss": 0.7102,
"step": 156
},
{
"epoch": 0.060038240917782024,
"grad_norm": 2.6535932830808338,
"learning_rate": 7.045800539528896e-07,
"loss": 0.2475,
"step": 157
},
{
"epoch": 0.060420650095602296,
"grad_norm": 2.724522743078367,
"learning_rate": 7.054648090365243e-07,
"loss": 0.4209,
"step": 158
},
{
"epoch": 0.06080305927342256,
"grad_norm": 3.0125584007079067,
"learning_rate": 7.063439820320076e-07,
"loss": 0.4966,
"step": 159
},
{
"epoch": 0.06118546845124283,
"grad_norm": 3.511451787422904,
"learning_rate": 7.072176429348676e-07,
"loss": 0.4125,
"step": 160
},
{
"epoch": 0.0615678776290631,
"grad_norm": 4.837683638209349,
"learning_rate": 7.080858604322876e-07,
"loss": 0.6034,
"step": 161
},
{
"epoch": 0.061950286806883366,
"grad_norm": 3.7945186383087077,
"learning_rate": 7.089487019355111e-07,
"loss": 0.3376,
"step": 162
},
{
"epoch": 0.06233269598470363,
"grad_norm": 7.647899074910034,
"learning_rate": 7.098062336112514e-07,
"loss": 0.6145,
"step": 163
},
{
"epoch": 0.0627151051625239,
"grad_norm": 6.282933251606341,
"learning_rate": 7.106585204121373e-07,
"loss": 1.0463,
"step": 164
},
{
"epoch": 0.06309751434034416,
"grad_norm": 3.282497350571109,
"learning_rate": 7.115056261062372e-07,
"loss": 0.5178,
"step": 165
},
{
"epoch": 0.06347992351816444,
"grad_norm": 2.1352796433069017,
"learning_rate": 7.123476133056884e-07,
"loss": 0.6417,
"step": 166
},
{
"epoch": 0.06386233269598471,
"grad_norm": 2.5930365994625575,
"learning_rate": 7.131845434944682e-07,
"loss": 0.4984,
"step": 167
},
{
"epoch": 0.06424474187380497,
"grad_norm": 3.0331021971899714,
"learning_rate": 7.140164770553345e-07,
"loss": 0.454,
"step": 168
},
{
"epoch": 0.06462715105162524,
"grad_norm": 3.7143123017921016,
"learning_rate": 7.148434732959689e-07,
"loss": 0.5334,
"step": 169
},
{
"epoch": 0.06500956022944551,
"grad_norm": 2.653427329658854,
"learning_rate": 7.156655904743474e-07,
"loss": 0.3391,
"step": 170
},
{
"epoch": 0.06539196940726577,
"grad_norm": 3.9509925367882985,
"learning_rate": 7.164828858233672e-07,
"loss": 0.3772,
"step": 171
},
{
"epoch": 0.06577437858508604,
"grad_norm": 3.1080221175458673,
"learning_rate": 7.172954155747545e-07,
"loss": 0.1742,
"step": 172
},
{
"epoch": 0.06615678776290632,
"grad_norm": 3.1603834827334754,
"learning_rate": 7.181032349822828e-07,
"loss": 0.2136,
"step": 173
},
{
"epoch": 0.06653919694072657,
"grad_norm": 3.798146729538179,
"learning_rate": 7.189063983443172e-07,
"loss": 0.1902,
"step": 174
},
{
"epoch": 0.06692160611854685,
"grad_norm": 4.068236540992522,
"learning_rate": 7.197049590257169e-07,
"loss": 0.3806,
"step": 175
},
{
"epoch": 0.0673040152963671,
"grad_norm": 6.4222483282794265,
"learning_rate": 7.204989694791108e-07,
"loss": 0.3978,
"step": 176
},
{
"epoch": 0.06768642447418738,
"grad_norm": 4.017960993317785,
"learning_rate": 7.212884812655726e-07,
"loss": 0.78,
"step": 177
},
{
"epoch": 0.06806883365200765,
"grad_norm": 4.100137168636647,
"learning_rate": 7.220735450747116e-07,
"loss": 0.7534,
"step": 178
},
{
"epoch": 0.06845124282982791,
"grad_norm": 3.1819592357927955,
"learning_rate": 7.228542107442019e-07,
"loss": 0.4387,
"step": 179
},
{
"epoch": 0.06883365200764818,
"grad_norm": 3.426584008354179,
"learning_rate": 7.236305272787674e-07,
"loss": 0.5703,
"step": 180
},
{
"epoch": 0.06921606118546846,
"grad_norm": 5.465640857252583,
"learning_rate": 7.244025428686407e-07,
"loss": 0.6831,
"step": 181
},
{
"epoch": 0.06959847036328871,
"grad_norm": 4.072806796700166,
"learning_rate": 7.25170304907513e-07,
"loss": 0.6685,
"step": 182
},
{
"epoch": 0.06998087954110899,
"grad_norm": 3.824467496019909,
"learning_rate": 7.259338600099944e-07,
"loss": 0.2409,
"step": 183
},
{
"epoch": 0.07036328871892926,
"grad_norm": 2.8699744709221924,
"learning_rate": 7.266932540285943e-07,
"loss": 0.2914,
"step": 184
},
{
"epoch": 0.07074569789674952,
"grad_norm": 3.6637638189913524,
"learning_rate": 7.274485320702463e-07,
"loss": 0.311,
"step": 185
},
{
"epoch": 0.07112810707456979,
"grad_norm": 4.0116479081805005,
"learning_rate": 7.281997385123849e-07,
"loss": 0.4676,
"step": 186
},
{
"epoch": 0.07151051625239006,
"grad_norm": 5.0535021280835455,
"learning_rate": 7.289469170185906e-07,
"loss": 0.2098,
"step": 187
},
{
"epoch": 0.07189292543021032,
"grad_norm": 5.6749955097466405,
"learning_rate": 7.29690110553821e-07,
"loss": 0.4254,
"step": 188
},
{
"epoch": 0.0722753346080306,
"grad_norm": 4.620593059176047,
"learning_rate": 7.304293613992342e-07,
"loss": 0.8025,
"step": 189
},
{
"epoch": 0.07265774378585087,
"grad_norm": 4.98679630219835,
"learning_rate": 7.311647111666231e-07,
"loss": 0.495,
"step": 190
},
{
"epoch": 0.07304015296367113,
"grad_norm": 3.0927893177777492,
"learning_rate": 7.318962008124717e-07,
"loss": 1.122,
"step": 191
},
{
"epoch": 0.0734225621414914,
"grad_norm": 3.338170753541131,
"learning_rate": 7.326238706516411e-07,
"loss": 0.6298,
"step": 192
},
{
"epoch": 0.07380497131931166,
"grad_norm": 3.7093543600843475,
"learning_rate": 7.333477603707038e-07,
"loss": 0.6464,
"step": 193
},
{
"epoch": 0.07418738049713193,
"grad_norm": 3.032196998144923,
"learning_rate": 7.340679090409308e-07,
"loss": 0.355,
"step": 194
},
{
"epoch": 0.0745697896749522,
"grad_norm": 2.8795885703469266,
"learning_rate": 7.34784355130946e-07,
"loss": 0.446,
"step": 195
},
{
"epoch": 0.07495219885277246,
"grad_norm": 2.5962306649778797,
"learning_rate": 7.354971365190572e-07,
"loss": 0.5733,
"step": 196
},
{
"epoch": 0.07533460803059273,
"grad_norm": 2.92991291450522,
"learning_rate": 7.36206290505272e-07,
"loss": 0.452,
"step": 197
},
{
"epoch": 0.075717017208413,
"grad_norm": 3.6905001057964744,
"learning_rate": 7.369118538230106e-07,
"loss": 0.4618,
"step": 198
},
{
"epoch": 0.07609942638623327,
"grad_norm": 3.9153593735892276,
"learning_rate": 7.376138626505238e-07,
"loss": 0.2443,
"step": 199
},
{
"epoch": 0.07648183556405354,
"grad_norm": 3.9085249723017563,
"learning_rate": 7.383123526220235e-07,
"loss": 0.3602,
"step": 200
},
{
"epoch": 0.07686424474187381,
"grad_norm": 7.705678891740636,
"learning_rate": 7.390073588385376e-07,
"loss": 0.5742,
"step": 201
},
{
"epoch": 0.07724665391969407,
"grad_norm": 4.761261490106676,
"learning_rate": 7.396989158784932e-07,
"loss": 0.4602,
"step": 202
},
{
"epoch": 0.07762906309751434,
"grad_norm": 3.166197708607704,
"learning_rate": 7.403870578080401e-07,
"loss": 0.3944,
"step": 203
},
{
"epoch": 0.07801147227533461,
"grad_norm": 2.0182306761377085,
"learning_rate": 7.410718181911208e-07,
"loss": 0.8064,
"step": 204
},
{
"epoch": 0.07839388145315487,
"grad_norm": 3.179968819135469,
"learning_rate": 7.417532300992931e-07,
"loss": 0.4025,
"step": 205
},
{
"epoch": 0.07877629063097515,
"grad_norm": 4.173314221331269,
"learning_rate": 7.424313261213144e-07,
"loss": 0.7782,
"step": 206
},
{
"epoch": 0.0791586998087954,
"grad_norm": 3.1765811377343054,
"learning_rate": 7.431061383724939e-07,
"loss": 0.6397,
"step": 207
},
{
"epoch": 0.07954110898661568,
"grad_norm": 4.6575972892850315,
"learning_rate": 7.437776985038198e-07,
"loss": 0.4831,
"step": 208
},
{
"epoch": 0.07992351816443595,
"grad_norm": 3.76169593304996,
"learning_rate": 7.444460377108663e-07,
"loss": 0.228,
"step": 209
},
{
"epoch": 0.08030592734225621,
"grad_norm": 3.9842518506933544,
"learning_rate": 7.451111867424903e-07,
"loss": 0.4997,
"step": 210
},
{
"epoch": 0.08068833652007648,
"grad_norm": 3.313081798203436,
"learning_rate": 7.457731759093186e-07,
"loss": 0.3388,
"step": 211
},
{
"epoch": 0.08107074569789675,
"grad_norm": 4.346696691194836,
"learning_rate": 7.464320350920372e-07,
"loss": 0.2379,
"step": 212
},
{
"epoch": 0.08145315487571701,
"grad_norm": 5.402567938113109,
"learning_rate": 7.470877937494842e-07,
"loss": 0.4411,
"step": 213
},
{
"epoch": 0.08183556405353729,
"grad_norm": 4.638690789435122,
"learning_rate": 7.477404809265524e-07,
"loss": 0.6799,
"step": 214
},
{
"epoch": 0.08221797323135756,
"grad_norm": 4.500483169047832,
"learning_rate": 7.483901252619104e-07,
"loss": 0.6008,
"step": 215
},
{
"epoch": 0.08260038240917782,
"grad_norm": 2.7526394575028728,
"learning_rate": 7.490367549955409e-07,
"loss": 0.4236,
"step": 216
},
{
"epoch": 0.08298279158699809,
"grad_norm": 2.745932171383271,
"learning_rate": 7.496803979761077e-07,
"loss": 0.2916,
"step": 217
},
{
"epoch": 0.08336520076481836,
"grad_norm": 3.1189073297420293,
"learning_rate": 7.50321081668153e-07,
"loss": 0.5905,
"step": 218
},
{
"epoch": 0.08374760994263862,
"grad_norm": 4.533831262357185,
"learning_rate": 7.509588331591282e-07,
"loss": 0.5721,
"step": 219
},
{
"epoch": 0.0841300191204589,
"grad_norm": 2.9649262548735305,
"learning_rate": 7.515936791662668e-07,
"loss": 0.3887,
"step": 220
},
{
"epoch": 0.08451242829827915,
"grad_norm": 3.4466325994562816,
"learning_rate": 7.522256460432995e-07,
"loss": 0.5528,
"step": 221
},
{
"epoch": 0.08489483747609942,
"grad_norm": 3.6307072928242223,
"learning_rate": 7.528547597870197e-07,
"loss": 0.5246,
"step": 222
},
{
"epoch": 0.0852772466539197,
"grad_norm": 3.257998381144178,
"learning_rate": 7.534810460436997e-07,
"loss": 0.2334,
"step": 223
},
{
"epoch": 0.08565965583173996,
"grad_norm": 4.644430635868781,
"learning_rate": 7.541045301153639e-07,
"loss": 0.2095,
"step": 224
},
{
"epoch": 0.08604206500956023,
"grad_norm": 4.69510308685718,
"learning_rate": 7.547252369659232e-07,
"loss": 0.3575,
"step": 225
},
{
"epoch": 0.0864244741873805,
"grad_norm": 6.285179920519442,
"learning_rate": 7.553431912271721e-07,
"loss": 0.6316,
"step": 226
},
{
"epoch": 0.08680688336520076,
"grad_norm": 4.806351337927113,
"learning_rate": 7.559584172046536e-07,
"loss": 0.6541,
"step": 227
},
{
"epoch": 0.08718929254302103,
"grad_norm": 3.001050606188733,
"learning_rate": 7.565709388833966e-07,
"loss": 0.7255,
"step": 228
},
{
"epoch": 0.0875717017208413,
"grad_norm": 1.9583878069937495,
"learning_rate": 7.571807799335257e-07,
"loss": 0.5502,
"step": 229
},
{
"epoch": 0.08795411089866156,
"grad_norm": 4.884711154296954,
"learning_rate": 7.577879637157502e-07,
"loss": 0.9388,
"step": 230
},
{
"epoch": 0.08833652007648184,
"grad_norm": 2.907146356646814,
"learning_rate": 7.583925132867335e-07,
"loss": 0.4075,
"step": 231
},
{
"epoch": 0.08871892925430211,
"grad_norm": 2.7039529050754956,
"learning_rate": 7.589944514043468e-07,
"loss": 0.3678,
"step": 232
},
{
"epoch": 0.08910133843212237,
"grad_norm": 5.727635219381453,
"learning_rate": 7.595938005328094e-07,
"loss": 0.5911,
"step": 233
},
{
"epoch": 0.08948374760994264,
"grad_norm": 4.664435648088752,
"learning_rate": 7.601905828477195e-07,
"loss": 0.2719,
"step": 234
},
{
"epoch": 0.08986615678776291,
"grad_norm": 3.5945420428990595,
"learning_rate": 7.607848202409769e-07,
"loss": 0.366,
"step": 235
},
{
"epoch": 0.09024856596558317,
"grad_norm": 4.234453826958854,
"learning_rate": 7.613765343256024e-07,
"loss": 0.2863,
"step": 236
},
{
"epoch": 0.09063097514340344,
"grad_norm": 7.158668076219786,
"learning_rate": 7.619657464404537e-07,
"loss": 0.1991,
"step": 237
},
{
"epoch": 0.0910133843212237,
"grad_norm": 7.641112766415768,
"learning_rate": 7.625524776548438e-07,
"loss": 0.5713,
"step": 238
},
{
"epoch": 0.09139579349904398,
"grad_norm": 4.9158103757560285,
"learning_rate": 7.631367487730614e-07,
"loss": 0.6582,
"step": 239
},
{
"epoch": 0.09177820267686425,
"grad_norm": 4.465435372685059,
"learning_rate": 7.63718580338797e-07,
"loss": 0.3755,
"step": 240
},
{
"epoch": 0.09216061185468451,
"grad_norm": 1.2563248360581727,
"learning_rate": 7.642979926394772e-07,
"loss": 0.8277,
"step": 241
},
{
"epoch": 0.09254302103250478,
"grad_norm": 2.627476840656689,
"learning_rate": 7.6487500571051e-07,
"loss": 0.4431,
"step": 242
},
{
"epoch": 0.09292543021032505,
"grad_norm": 3.4558729742159238,
"learning_rate": 7.654496393394404e-07,
"loss": 0.7704,
"step": 243
},
{
"epoch": 0.09330783938814531,
"grad_norm": 2.711048310843166,
"learning_rate": 7.660219130700239e-07,
"loss": 0.3752,
"step": 244
},
{
"epoch": 0.09369024856596558,
"grad_norm": 2.259893705518403,
"learning_rate": 7.665918462062132e-07,
"loss": 0.4386,
"step": 245
},
{
"epoch": 0.09407265774378586,
"grad_norm": 2.9605598309499297,
"learning_rate": 7.671594578160665e-07,
"loss": 0.356,
"step": 246
},
{
"epoch": 0.09445506692160612,
"grad_norm": 4.510125953285779,
"learning_rate": 7.677247667355753e-07,
"loss": 0.3476,
"step": 247
},
{
"epoch": 0.09483747609942639,
"grad_norm": 4.195827573275409,
"learning_rate": 7.682877915724144e-07,
"loss": 0.3094,
"step": 248
},
{
"epoch": 0.09521988527724666,
"grad_norm": 3.3973305488866457,
"learning_rate": 7.688485507096177e-07,
"loss": 0.1854,
"step": 249
},
{
"epoch": 0.09560229445506692,
"grad_norm": 6.474794374242375,
"learning_rate": 7.694070623091794e-07,
"loss": 0.3838,
"step": 250
},
{
"epoch": 0.09598470363288719,
"grad_norm": 5.565976203234317,
"learning_rate": 7.699633443155836e-07,
"loss": 0.4187,
"step": 251
},
{
"epoch": 0.09636711281070745,
"grad_norm": 4.445144730605,
"learning_rate": 7.705174144592636e-07,
"loss": 0.5566,
"step": 252
},
{
"epoch": 0.09674952198852772,
"grad_norm": 4.125839841484762,
"learning_rate": 7.710692902599934e-07,
"loss": 0.3737,
"step": 253
},
{
"epoch": 0.097131931166348,
"grad_norm": 6.788624212693753,
"learning_rate": 7.716189890302107e-07,
"loss": 0.6117,
"step": 254
},
{
"epoch": 0.09751434034416825,
"grad_norm": 3.3993700173101784,
"learning_rate": 7.721665278782767e-07,
"loss": 0.5299,
"step": 255
},
{
"epoch": 0.09789674952198853,
"grad_norm": 3.5447109527471192,
"learning_rate": 7.727119237116706e-07,
"loss": 0.464,
"step": 256
},
{
"epoch": 0.0982791586998088,
"grad_norm": 3.451375250445069,
"learning_rate": 7.732551932401223e-07,
"loss": 0.3774,
"step": 257
},
{
"epoch": 0.09866156787762906,
"grad_norm": 2.985844594806625,
"learning_rate": 7.737963529786837e-07,
"loss": 0.4458,
"step": 258
},
{
"epoch": 0.09904397705544933,
"grad_norm": 3.433412792310995,
"learning_rate": 7.743354192507426e-07,
"loss": 0.4216,
"step": 259
},
{
"epoch": 0.0994263862332696,
"grad_norm": 3.1338503823562065,
"learning_rate": 7.748724081909757e-07,
"loss": 0.3232,
"step": 260
},
{
"epoch": 0.09980879541108986,
"grad_norm": 4.060734526613366,
"learning_rate": 7.754073357482466e-07,
"loss": 0.4153,
"step": 261
},
{
"epoch": 0.10019120458891013,
"grad_norm": 3.565755460928457,
"learning_rate": 7.759402176884483e-07,
"loss": 0.2684,
"step": 262
},
{
"epoch": 0.10057361376673041,
"grad_norm": 5.585943816921691,
"learning_rate": 7.764710695972918e-07,
"loss": 0.5857,
"step": 263
},
{
"epoch": 0.10095602294455067,
"grad_norm": 4.169284659304579,
"learning_rate": 7.769999068830402e-07,
"loss": 1.0044,
"step": 264
},
{
"epoch": 0.10133843212237094,
"grad_norm": 4.71342197779094,
"learning_rate": 7.775267447791931e-07,
"loss": 0.7013,
"step": 265
},
{
"epoch": 0.10172084130019121,
"grad_norm": 1.434724367100525,
"learning_rate": 7.780515983471195e-07,
"loss": 0.6375,
"step": 266
},
{
"epoch": 0.10210325047801147,
"grad_norm": 3.895493490862681,
"learning_rate": 7.785744824786409e-07,
"loss": 0.4048,
"step": 267
},
{
"epoch": 0.10248565965583174,
"grad_norm": 3.8662095808203607,
"learning_rate": 7.790954118985672e-07,
"loss": 0.7465,
"step": 268
},
{
"epoch": 0.102868068833652,
"grad_norm": 3.424467036971547,
"learning_rate": 7.796144011671844e-07,
"loss": 0.408,
"step": 269
},
{
"epoch": 0.10325047801147227,
"grad_norm": 3.550724310255783,
"learning_rate": 7.801314646826968e-07,
"loss": 0.3537,
"step": 270
},
{
"epoch": 0.10363288718929255,
"grad_norm": 2.89688389518549,
"learning_rate": 7.80646616683623e-07,
"loss": 0.3399,
"step": 271
},
{
"epoch": 0.1040152963671128,
"grad_norm": 4.119613367105412,
"learning_rate": 7.811598712511505e-07,
"loss": 0.3759,
"step": 272
},
{
"epoch": 0.10439770554493308,
"grad_norm": 2.973462029424571,
"learning_rate": 7.816712423114423e-07,
"loss": 0.2155,
"step": 273
},
{
"epoch": 0.10478011472275335,
"grad_norm": 3.3977701167614947,
"learning_rate": 7.82180743637907e-07,
"loss": 0.2368,
"step": 274
},
{
"epoch": 0.10516252390057361,
"grad_norm": 4.181511767850576,
"learning_rate": 7.826883888534227e-07,
"loss": 0.2863,
"step": 275
},
{
"epoch": 0.10554493307839388,
"grad_norm": 5.954453848161291,
"learning_rate": 7.831941914325235e-07,
"loss": 0.622,
"step": 276
},
{
"epoch": 0.10592734225621415,
"grad_norm": 4.171188698084672,
"learning_rate": 7.836981647035458e-07,
"loss": 0.9351,
"step": 277
},
{
"epoch": 0.10630975143403441,
"grad_norm": 2.846374809467165,
"learning_rate": 7.842003218507345e-07,
"loss": 0.6559,
"step": 278
},
{
"epoch": 0.10669216061185469,
"grad_norm": 2.124035560310232,
"learning_rate": 7.847006759163141e-07,
"loss": 0.5626,
"step": 279
},
{
"epoch": 0.10707456978967496,
"grad_norm": 3.5632323466003974,
"learning_rate": 7.851992398025198e-07,
"loss": 0.3027,
"step": 280
},
{
"epoch": 0.10745697896749522,
"grad_norm": 2.80981512133499,
"learning_rate": 7.856960262735946e-07,
"loss": 0.4338,
"step": 281
},
{
"epoch": 0.10783938814531549,
"grad_norm": 3.1493700033255347,
"learning_rate": 7.861910479577502e-07,
"loss": 0.3888,
"step": 282
},
{
"epoch": 0.10822179732313575,
"grad_norm": 3.199839325685757,
"learning_rate": 7.866843173490923e-07,
"loss": 0.581,
"step": 283
},
{
"epoch": 0.10860420650095602,
"grad_norm": 3.9744033797586047,
"learning_rate": 7.871758468095137e-07,
"loss": 0.3301,
"step": 284
},
{
"epoch": 0.1089866156787763,
"grad_norm": 4.082176563651574,
"learning_rate": 7.876656485705525e-07,
"loss": 0.2453,
"step": 285
},
{
"epoch": 0.10936902485659655,
"grad_norm": 3.762870572619933,
"learning_rate": 7.881537347352188e-07,
"loss": 0.1672,
"step": 286
},
{
"epoch": 0.10975143403441683,
"grad_norm": 4.497364119915088,
"learning_rate": 7.886401172797894e-07,
"loss": 0.1229,
"step": 287
},
{
"epoch": 0.1101338432122371,
"grad_norm": 6.455440373723637,
"learning_rate": 7.891248080555703e-07,
"loss": 0.2936,
"step": 288
},
{
"epoch": 0.11051625239005736,
"grad_norm": 3.5754464464259237,
"learning_rate": 7.896078187906302e-07,
"loss": 0.7403,
"step": 289
},
{
"epoch": 0.11089866156787763,
"grad_norm": 2.764753991163198,
"learning_rate": 7.900891610915027e-07,
"loss": 0.5118,
"step": 290
},
{
"epoch": 0.1112810707456979,
"grad_norm": 2.297678385254701,
"learning_rate": 7.9056884644486e-07,
"loss": 0.4936,
"step": 291
},
{
"epoch": 0.11166347992351816,
"grad_norm": 1.8158157537320423,
"learning_rate": 7.910468862191579e-07,
"loss": 0.7025,
"step": 292
},
{
"epoch": 0.11204588910133843,
"grad_norm": 3.552000535165901,
"learning_rate": 7.915232916662518e-07,
"loss": 0.706,
"step": 293
},
{
"epoch": 0.1124282982791587,
"grad_norm": 3.6530113548438585,
"learning_rate": 7.919980739229866e-07,
"loss": 0.3169,
"step": 294
},
{
"epoch": 0.11281070745697896,
"grad_norm": 3.2632719624729427,
"learning_rate": 7.924712440127582e-07,
"loss": 0.3265,
"step": 295
},
{
"epoch": 0.11319311663479924,
"grad_norm": 3.110373025338036,
"learning_rate": 7.929428128470493e-07,
"loss": 0.3102,
"step": 296
},
{
"epoch": 0.1135755258126195,
"grad_norm": 3.6801332499387422,
"learning_rate": 7.9341279122694e-07,
"loss": 0.423,
"step": 297
},
{
"epoch": 0.11395793499043977,
"grad_norm": 3.8900937862907234,
"learning_rate": 7.938811898445904e-07,
"loss": 0.2004,
"step": 298
},
{
"epoch": 0.11434034416826004,
"grad_norm": 3.7375127785760855,
"learning_rate": 7.943480192847023e-07,
"loss": 0.1989,
"step": 299
},
{
"epoch": 0.1147227533460803,
"grad_norm": 4.461046741971845,
"learning_rate": 7.948132900259528e-07,
"loss": 0.3909,
"step": 300
},
{
"epoch": 0.11510516252390057,
"grad_norm": 6.173622057914207,
"learning_rate": 7.952770124424067e-07,
"loss": 0.3894,
"step": 301
},
{
"epoch": 0.11548757170172085,
"grad_norm": 3.733296769394523,
"learning_rate": 7.957391968049033e-07,
"loss": 0.8673,
"step": 302
},
{
"epoch": 0.1158699808795411,
"grad_norm": 3.9489473123211307,
"learning_rate": 7.961998532824224e-07,
"loss": 0.6133,
"step": 303
},
{
"epoch": 0.11625239005736138,
"grad_norm": 4.475476091515576,
"learning_rate": 7.966589919434262e-07,
"loss": 0.7168,
"step": 304
},
{
"epoch": 0.11663479923518165,
"grad_norm": 2.3809739320778744,
"learning_rate": 7.971166227571798e-07,
"loss": 0.5665,
"step": 305
},
{
"epoch": 0.11701720841300191,
"grad_norm": 3.904736156855016,
"learning_rate": 7.975727555950501e-07,
"loss": 0.4922,
"step": 306
},
{
"epoch": 0.11739961759082218,
"grad_norm": 3.4877422125598634,
"learning_rate": 7.980274002317836e-07,
"loss": 0.3931,
"step": 307
},
{
"epoch": 0.11778202676864245,
"grad_norm": 3.089792822648413,
"learning_rate": 7.98480566346763e-07,
"loss": 0.3242,
"step": 308
},
{
"epoch": 0.11816443594646271,
"grad_norm": 3.1274358792520442,
"learning_rate": 7.989322635252437e-07,
"loss": 0.2631,
"step": 309
},
{
"epoch": 0.11854684512428298,
"grad_norm": 3.306093606693292,
"learning_rate": 7.993825012595703e-07,
"loss": 0.1887,
"step": 310
},
{
"epoch": 0.11892925430210326,
"grad_norm": 5.434712364074534,
"learning_rate": 7.998312889503738e-07,
"loss": 0.2773,
"step": 311
},
{
"epoch": 0.11931166347992352,
"grad_norm": 4.773654664339484,
"learning_rate": 8.00278635907749e-07,
"loss": 0.1587,
"step": 312
},
{
"epoch": 0.11969407265774379,
"grad_norm": 5.61964161283346,
"learning_rate": 8.007245513524136e-07,
"loss": 0.4011,
"step": 313
},
{
"epoch": 0.12007648183556405,
"grad_norm": 5.422191951403171,
"learning_rate": 8.011690444168484e-07,
"loss": 0.5186,
"step": 314
},
{
"epoch": 0.12045889101338432,
"grad_norm": 4.235650240037841,
"learning_rate": 8.016121241464195e-07,
"loss": 0.6452,
"step": 315
},
{
"epoch": 0.12084130019120459,
"grad_norm": 2.2959913772859717,
"learning_rate": 8.020537995004831e-07,
"loss": 0.3568,
"step": 316
},
{
"epoch": 0.12122370936902485,
"grad_norm": 2.7906570791555816,
"learning_rate": 8.024940793534721e-07,
"loss": 0.5597,
"step": 317
},
{
"epoch": 0.12160611854684512,
"grad_norm": 3.0943336321631882,
"learning_rate": 8.029329724959665e-07,
"loss": 0.5461,
"step": 318
},
{
"epoch": 0.1219885277246654,
"grad_norm": 3.4053237075530216,
"learning_rate": 8.033704876357459e-07,
"loss": 0.4284,
"step": 319
},
{
"epoch": 0.12237093690248566,
"grad_norm": 3.234592596281024,
"learning_rate": 8.038066333988265e-07,
"loss": 0.6019,
"step": 320
},
{
"epoch": 0.12275334608030593,
"grad_norm": 3.4472462596835673,
"learning_rate": 8.042414183304817e-07,
"loss": 0.5452,
"step": 321
},
{
"epoch": 0.1231357552581262,
"grad_norm": 3.6241358022865984,
"learning_rate": 8.046748508962464e-07,
"loss": 0.3125,
"step": 322
},
{
"epoch": 0.12351816443594646,
"grad_norm": 3.893943260136077,
"learning_rate": 8.051069394829059e-07,
"loss": 0.2509,
"step": 323
},
{
"epoch": 0.12390057361376673,
"grad_norm": 4.8189353026547135,
"learning_rate": 8.0553769239947e-07,
"loss": 0.1644,
"step": 324
},
{
"epoch": 0.124282982791587,
"grad_norm": 3.559991991055513,
"learning_rate": 8.059671178781316e-07,
"loss": 0.3537,
"step": 325
},
{
"epoch": 0.12466539196940726,
"grad_norm": 6.39953827384509,
"learning_rate": 8.063952240752102e-07,
"loss": 0.5631,
"step": 326
},
{
"epoch": 0.12504780114722752,
"grad_norm": 3.6130294464626025,
"learning_rate": 8.068220190720823e-07,
"loss": 0.7063,
"step": 327
},
{
"epoch": 0.1254302103250478,
"grad_norm": 3.1183942022317663,
"learning_rate": 8.072475108760961e-07,
"loss": 0.7305,
"step": 328
},
{
"epoch": 0.12581261950286807,
"grad_norm": 2.745672983244802,
"learning_rate": 8.076717074214732e-07,
"loss": 0.4498,
"step": 329
},
{
"epoch": 0.12619502868068833,
"grad_norm": 2.8516755382133048,
"learning_rate": 8.08094616570196e-07,
"loss": 0.356,
"step": 330
},
{
"epoch": 0.1265774378585086,
"grad_norm": 3.0513196560458202,
"learning_rate": 8.085162461128826e-07,
"loss": 0.2682,
"step": 331
},
{
"epoch": 0.12695984703632887,
"grad_norm": 4.121776361564982,
"learning_rate": 8.089366037696473e-07,
"loss": 0.4183,
"step": 332
},
{
"epoch": 0.12734225621414913,
"grad_norm": 2.9887883173753043,
"learning_rate": 8.093556971909491e-07,
"loss": 0.3487,
"step": 333
},
{
"epoch": 0.12772466539196942,
"grad_norm": 3.216439019351985,
"learning_rate": 8.09773533958427e-07,
"loss": 0.3038,
"step": 334
},
{
"epoch": 0.12810707456978968,
"grad_norm": 4.0139201660580275,
"learning_rate": 8.101901215857231e-07,
"loss": 0.2941,
"step": 335
},
{
"epoch": 0.12848948374760993,
"grad_norm": 3.350848924483089,
"learning_rate": 8.106054675192933e-07,
"loss": 0.175,
"step": 336
},
{
"epoch": 0.12887189292543022,
"grad_norm": 3.438037899587354,
"learning_rate": 8.110195791392056e-07,
"loss": 0.1396,
"step": 337
},
{
"epoch": 0.12925430210325048,
"grad_norm": 6.772774599013537,
"learning_rate": 8.114324637599278e-07,
"loss": 0.5966,
"step": 338
},
{
"epoch": 0.12963671128107074,
"grad_norm": 4.137346920627515,
"learning_rate": 8.118441286311013e-07,
"loss": 0.4748,
"step": 339
},
{
"epoch": 0.13001912045889102,
"grad_norm": 3.4568218620380566,
"learning_rate": 8.122545809383064e-07,
"loss": 0.5865,
"step": 340
},
{
"epoch": 0.13040152963671128,
"grad_norm": 3.0236478825988216,
"learning_rate": 8.126638278038134e-07,
"loss": 0.4805,
"step": 341
},
{
"epoch": 0.13078393881453154,
"grad_norm": 2.559726796175156,
"learning_rate": 8.130718762873258e-07,
"loss": 0.3147,
"step": 342
},
{
"epoch": 0.13116634799235183,
"grad_norm": 4.449152087749615,
"learning_rate": 8.134787333867094e-07,
"loss": 0.9307,
"step": 343
},
{
"epoch": 0.1315487571701721,
"grad_norm": 2.6546079160044753,
"learning_rate": 8.138844060387134e-07,
"loss": 0.2215,
"step": 344
},
{
"epoch": 0.13193116634799235,
"grad_norm": 2.47569283700409,
"learning_rate": 8.142889011196796e-07,
"loss": 0.4161,
"step": 345
},
{
"epoch": 0.13231357552581263,
"grad_norm": 4.2666821409978,
"learning_rate": 8.146922254462416e-07,
"loss": 0.2177,
"step": 346
},
{
"epoch": 0.1326959847036329,
"grad_norm": 3.236865425515548,
"learning_rate": 8.150943857760151e-07,
"loss": 0.2703,
"step": 347
},
{
"epoch": 0.13307839388145315,
"grad_norm": 4.03876947339817,
"learning_rate": 8.154953888082761e-07,
"loss": 0.4093,
"step": 348
},
{
"epoch": 0.13346080305927344,
"grad_norm": 5.638943831119226,
"learning_rate": 8.158952411846307e-07,
"loss": 0.3396,
"step": 349
},
{
"epoch": 0.1338432122370937,
"grad_norm": 4.114233821397829,
"learning_rate": 8.162939494896757e-07,
"loss": 0.346,
"step": 350
},
{
"epoch": 0.13422562141491395,
"grad_norm": 7.785144000084442,
"learning_rate": 8.166915202516487e-07,
"loss": 0.4865,
"step": 351
},
{
"epoch": 0.1346080305927342,
"grad_norm": 4.286640643865958,
"learning_rate": 8.170879599430697e-07,
"loss": 0.6353,
"step": 352
},
{
"epoch": 0.1349904397705545,
"grad_norm": 3.460942184410628,
"learning_rate": 8.174832749813731e-07,
"loss": 0.5312,
"step": 353
},
{
"epoch": 0.13537284894837476,
"grad_norm": 2.148699158427092,
"learning_rate": 8.178774717295314e-07,
"loss": 0.3499,
"step": 354
},
{
"epoch": 0.13575525812619502,
"grad_norm": 2.5779189518750436,
"learning_rate": 8.182705564966695e-07,
"loss": 0.3294,
"step": 355
},
{
"epoch": 0.1361376673040153,
"grad_norm": 3.7108314900565365,
"learning_rate": 8.186625355386705e-07,
"loss": 0.5441,
"step": 356
},
{
"epoch": 0.13652007648183556,
"grad_norm": 2.8954350222722463,
"learning_rate": 8.19053415058773e-07,
"loss": 0.4693,
"step": 357
},
{
"epoch": 0.13690248565965582,
"grad_norm": 3.8289599695272587,
"learning_rate": 8.194432012081608e-07,
"loss": 0.3562,
"step": 358
},
{
"epoch": 0.1372848948374761,
"grad_norm": 4.809977873753277,
"learning_rate": 8.198319000865427e-07,
"loss": 0.4452,
"step": 359
},
{
"epoch": 0.13766730401529637,
"grad_norm": 3.428323322356324,
"learning_rate": 8.202195177427262e-07,
"loss": 0.3554,
"step": 360
},
{
"epoch": 0.13804971319311662,
"grad_norm": 6.486314304317555,
"learning_rate": 8.206060601751816e-07,
"loss": 0.1975,
"step": 361
},
{
"epoch": 0.1384321223709369,
"grad_norm": 3.8071659483116043,
"learning_rate": 8.209915333325995e-07,
"loss": 0.2016,
"step": 362
},
{
"epoch": 0.13881453154875717,
"grad_norm": 7.472006881743372,
"learning_rate": 8.213759431144391e-07,
"loss": 0.4247,
"step": 363
},
{
"epoch": 0.13919694072657743,
"grad_norm": 4.879357371307577,
"learning_rate": 8.21759295371472e-07,
"loss": 0.6671,
"step": 364
},
{
"epoch": 0.13957934990439771,
"grad_norm": 5.342405808242789,
"learning_rate": 8.221415959063138e-07,
"loss": 0.5423,
"step": 365
},
{
"epoch": 0.13996175908221797,
"grad_norm": 4.174753631815001,
"learning_rate": 8.225228504739531e-07,
"loss": 0.269,
"step": 366
},
{
"epoch": 0.14034416826003823,
"grad_norm": 3.108287678673902,
"learning_rate": 8.229030647822711e-07,
"loss": 0.6411,
"step": 367
},
{
"epoch": 0.14072657743785852,
"grad_norm": 3.4124323184948455,
"learning_rate": 8.232822444925531e-07,
"loss": 0.4184,
"step": 368
},
{
"epoch": 0.14110898661567878,
"grad_norm": 3.5585165674944963,
"learning_rate": 8.236603952199958e-07,
"loss": 0.6019,
"step": 369
},
{
"epoch": 0.14149139579349904,
"grad_norm": 4.025666621193951,
"learning_rate": 8.240375225342052e-07,
"loss": 0.4895,
"step": 370
},
{
"epoch": 0.14187380497131932,
"grad_norm": 4.571471082115062,
"learning_rate": 8.244136319596894e-07,
"loss": 0.2911,
"step": 371
},
{
"epoch": 0.14225621414913958,
"grad_norm": 2.5329872104037254,
"learning_rate": 8.247887289763437e-07,
"loss": 0.205,
"step": 372
},
{
"epoch": 0.14263862332695984,
"grad_norm": 4.108219114705737,
"learning_rate": 8.251628190199301e-07,
"loss": 0.2088,
"step": 373
},
{
"epoch": 0.14302103250478013,
"grad_norm": 3.533705541211625,
"learning_rate": 8.255359074825496e-07,
"loss": 0.2682,
"step": 374
},
{
"epoch": 0.14340344168260039,
"grad_norm": 4.903078840243899,
"learning_rate": 8.259079997131087e-07,
"loss": 0.2899,
"step": 375
},
{
"epoch": 0.14378585086042064,
"grad_norm": 7.337460165874226,
"learning_rate": 8.262791010177799e-07,
"loss": 0.5831,
"step": 376
},
{
"epoch": 0.14416826003824093,
"grad_norm": 3.68396977986088,
"learning_rate": 8.266492166604547e-07,
"loss": 0.7252,
"step": 377
},
{
"epoch": 0.1445506692160612,
"grad_norm": 3.7188438596020146,
"learning_rate": 8.27018351863193e-07,
"loss": 0.473,
"step": 378
},
{
"epoch": 0.14493307839388145,
"grad_norm": 1.4870233300950009,
"learning_rate": 8.273865118066637e-07,
"loss": 0.5603,
"step": 379
},
{
"epoch": 0.14531548757170173,
"grad_norm": 2.2572929022034094,
"learning_rate": 8.27753701630582e-07,
"loss": 0.3373,
"step": 380
},
{
"epoch": 0.145697896749522,
"grad_norm": 3.0990770504221294,
"learning_rate": 8.2811992643414e-07,
"loss": 0.4191,
"step": 381
},
{
"epoch": 0.14608030592734225,
"grad_norm": 4.983702993982142,
"learning_rate": 8.284851912764304e-07,
"loss": 0.5062,
"step": 382
},
{
"epoch": 0.1464627151051625,
"grad_norm": 2.7700035650182153,
"learning_rate": 8.288495011768676e-07,
"loss": 0.3522,
"step": 383
},
{
"epoch": 0.1468451242829828,
"grad_norm": 2.503238963986765,
"learning_rate": 8.292128611156e-07,
"loss": 0.2989,
"step": 384
},
{
"epoch": 0.14722753346080306,
"grad_norm": 4.035654426834888,
"learning_rate": 8.29575276033919e-07,
"loss": 0.2093,
"step": 385
},
{
"epoch": 0.14760994263862331,
"grad_norm": 3.433500008719308,
"learning_rate": 8.299367508346627e-07,
"loss": 0.1824,
"step": 386
},
{
"epoch": 0.1479923518164436,
"grad_norm": 4.621602021590747,
"learning_rate": 8.302972903826131e-07,
"loss": 0.1898,
"step": 387
},
{
"epoch": 0.14837476099426386,
"grad_norm": 7.1558155444335645,
"learning_rate": 8.306568995048896e-07,
"loss": 0.5661,
"step": 388
},
{
"epoch": 0.14875717017208412,
"grad_norm": 4.818200438514181,
"learning_rate": 8.310155829913362e-07,
"loss": 0.7258,
"step": 389
},
{
"epoch": 0.1491395793499044,
"grad_norm": 4.729636663309238,
"learning_rate": 8.313733455949049e-07,
"loss": 0.7956,
"step": 390
},
{
"epoch": 0.14952198852772466,
"grad_norm": 4.556495373140328,
"learning_rate": 8.317301920320328e-07,
"loss": 0.4644,
"step": 391
},
{
"epoch": 0.14990439770554492,
"grad_norm": 2.829353529349588,
"learning_rate": 8.32086126983016e-07,
"loss": 0.3368,
"step": 392
},
{
"epoch": 0.1502868068833652,
"grad_norm": 2.855775662691332,
"learning_rate": 8.324411550923776e-07,
"loss": 0.445,
"step": 393
},
{
"epoch": 0.15066921606118547,
"grad_norm": 3.1955485734729185,
"learning_rate": 8.327952809692308e-07,
"loss": 0.4805,
"step": 394
},
{
"epoch": 0.15105162523900573,
"grad_norm": 2.628568675797958,
"learning_rate": 8.33148509187639e-07,
"loss": 0.3455,
"step": 395
},
{
"epoch": 0.151434034416826,
"grad_norm": 4.427946741019882,
"learning_rate": 8.335008442869694e-07,
"loss": 0.2197,
"step": 396
},
{
"epoch": 0.15181644359464627,
"grad_norm": 4.976054454603633,
"learning_rate": 8.338522907722437e-07,
"loss": 0.569,
"step": 397
},
{
"epoch": 0.15219885277246653,
"grad_norm": 3.0160754564089474,
"learning_rate": 8.342028531144826e-07,
"loss": 0.3384,
"step": 398
},
{
"epoch": 0.15258126195028682,
"grad_norm": 3.001475765915238,
"learning_rate": 8.345525357510488e-07,
"loss": 0.2957,
"step": 399
},
{
"epoch": 0.15296367112810708,
"grad_norm": 4.44835313476485,
"learning_rate": 8.349013430859824e-07,
"loss": 0.3145,
"step": 400
},
{
"epoch": 0.15334608030592733,
"grad_norm": 5.9358016948221834,
"learning_rate": 8.352492794903348e-07,
"loss": 0.5199,
"step": 401
},
{
"epoch": 0.15372848948374762,
"grad_norm": 3.8760502530236685,
"learning_rate": 8.355963493024964e-07,
"loss": 0.7258,
"step": 402
},
{
"epoch": 0.15411089866156788,
"grad_norm": 2.653734339821923,
"learning_rate": 8.359425568285223e-07,
"loss": 0.467,
"step": 403
},
{
"epoch": 0.15449330783938814,
"grad_norm": 3.2408897652245,
"learning_rate": 8.36287906342452e-07,
"loss": 0.5655,
"step": 404
},
{
"epoch": 0.15487571701720843,
"grad_norm": 5.064740097442185,
"learning_rate": 8.366324020866259e-07,
"loss": 0.4984,
"step": 405
},
{
"epoch": 0.15525812619502868,
"grad_norm": 4.151100582660372,
"learning_rate": 8.369760482719989e-07,
"loss": 0.3415,
"step": 406
},
{
"epoch": 0.15564053537284894,
"grad_norm": 2.1285019236890412,
"learning_rate": 8.373188490784484e-07,
"loss": 0.5143,
"step": 407
},
{
"epoch": 0.15602294455066923,
"grad_norm": 3.038817564991084,
"learning_rate": 8.376608086550797e-07,
"loss": 0.3227,
"step": 408
},
{
"epoch": 0.1564053537284895,
"grad_norm": 3.10377849908374,
"learning_rate": 8.380019311205272e-07,
"loss": 0.2325,
"step": 409
},
{
"epoch": 0.15678776290630975,
"grad_norm": 2.9242790246530648,
"learning_rate": 8.38342220563252e-07,
"loss": 0.2834,
"step": 410
},
{
"epoch": 0.15717017208413,
"grad_norm": 4.360945223969221,
"learning_rate": 8.386816810418363e-07,
"loss": 0.1797,
"step": 411
},
{
"epoch": 0.1575525812619503,
"grad_norm": 4.194239644376321,
"learning_rate": 8.390203165852732e-07,
"loss": 0.1658,
"step": 412
},
{
"epoch": 0.15793499043977055,
"grad_norm": 5.705579924332777,
"learning_rate": 8.393581311932545e-07,
"loss": 0.3086,
"step": 413
},
{
"epoch": 0.1583173996175908,
"grad_norm": 5.0412789608184765,
"learning_rate": 8.396951288364528e-07,
"loss": 0.5928,
"step": 414
},
{
"epoch": 0.1586998087954111,
"grad_norm": 4.067400983790442,
"learning_rate": 8.400313134568032e-07,
"loss": 0.4792,
"step": 415
},
{
"epoch": 0.15908221797323135,
"grad_norm": 2.325920546220163,
"learning_rate": 8.403666889677787e-07,
"loss": 0.3938,
"step": 416
},
{
"epoch": 0.1594646271510516,
"grad_norm": 2.529085434725153,
"learning_rate": 8.407012592546638e-07,
"loss": 0.4017,
"step": 417
},
{
"epoch": 0.1598470363288719,
"grad_norm": 4.022784667151783,
"learning_rate": 8.410350281748251e-07,
"loss": 0.4667,
"step": 418
},
{
"epoch": 0.16022944550669216,
"grad_norm": 5.081691581053613,
"learning_rate": 8.413679995579778e-07,
"loss": 0.3841,
"step": 419
},
{
"epoch": 0.16061185468451242,
"grad_norm": 2.8149274734721326,
"learning_rate": 8.417001772064491e-07,
"loss": 0.1482,
"step": 420
},
{
"epoch": 0.1609942638623327,
"grad_norm": 2.5876473942299856,
"learning_rate": 8.420315648954389e-07,
"loss": 0.219,
"step": 421
},
{
"epoch": 0.16137667304015296,
"grad_norm": 4.121041450349419,
"learning_rate": 8.423621663732774e-07,
"loss": 0.3829,
"step": 422
},
{
"epoch": 0.16175908221797322,
"grad_norm": 3.8963933260039,
"learning_rate": 8.426919853616795e-07,
"loss": 0.1564,
"step": 423
},
{
"epoch": 0.1621414913957935,
"grad_norm": 4.824550058208553,
"learning_rate": 8.430210255559961e-07,
"loss": 0.3041,
"step": 424
},
{
"epoch": 0.16252390057361377,
"grad_norm": 4.454444020661066,
"learning_rate": 8.433492906254623e-07,
"loss": 0.2036,
"step": 425
},
{
"epoch": 0.16290630975143403,
"grad_norm": 8.36002388690657,
"learning_rate": 8.43676784213443e-07,
"loss": 0.5876,
"step": 426
},
{
"epoch": 0.1632887189292543,
"grad_norm": 3.929175887732475,
"learning_rate": 8.440035099376761e-07,
"loss": 0.4755,
"step": 427
},
{
"epoch": 0.16367112810707457,
"grad_norm": 3.6463606680107525,
"learning_rate": 8.443294713905113e-07,
"loss": 0.5675,
"step": 428
},
{
"epoch": 0.16405353728489483,
"grad_norm": 3.6404536784059127,
"learning_rate": 8.446546721391481e-07,
"loss": 0.4856,
"step": 429
},
{
"epoch": 0.16443594646271512,
"grad_norm": 2.6217796917621112,
"learning_rate": 8.449791157258692e-07,
"loss": 0.6915,
"step": 430
},
{
"epoch": 0.16481835564053537,
"grad_norm": 3.42014551221246,
"learning_rate": 8.453028056682724e-07,
"loss": 0.627,
"step": 431
},
{
"epoch": 0.16520076481835563,
"grad_norm": 4.183283262725723,
"learning_rate": 8.456257454594997e-07,
"loss": 0.2805,
"step": 432
},
{
"epoch": 0.16558317399617592,
"grad_norm": 3.93192174657836,
"learning_rate": 8.459479385684625e-07,
"loss": 0.3485,
"step": 433
},
{
"epoch": 0.16596558317399618,
"grad_norm": 2.567881817531444,
"learning_rate": 8.462693884400665e-07,
"loss": 0.2407,
"step": 434
},
{
"epoch": 0.16634799235181644,
"grad_norm": 3.8513236755237674,
"learning_rate": 8.46590098495432e-07,
"loss": 0.1689,
"step": 435
},
{
"epoch": 0.16673040152963672,
"grad_norm": 5.43608618585279,
"learning_rate": 8.469100721321118e-07,
"loss": 0.3796,
"step": 436
},
{
"epoch": 0.16711281070745698,
"grad_norm": 3.427005492527403,
"learning_rate": 8.472293127243085e-07,
"loss": 0.1437,
"step": 437
},
{
"epoch": 0.16749521988527724,
"grad_norm": 7.971012527304747,
"learning_rate": 8.47547823623087e-07,
"loss": 0.4998,
"step": 438
},
{
"epoch": 0.16787762906309753,
"grad_norm": 4.308472231329761,
"learning_rate": 8.478656081565858e-07,
"loss": 0.6339,
"step": 439
},
{
"epoch": 0.1682600382409178,
"grad_norm": 4.791877255442337,
"learning_rate": 8.481826696302256e-07,
"loss": 0.6164,
"step": 440
},
{
"epoch": 0.16864244741873805,
"grad_norm": 3.8886846276787375,
"learning_rate": 8.484990113269158e-07,
"loss": 0.7065,
"step": 441
},
{
"epoch": 0.1690248565965583,
"grad_norm": 2.3271527404361185,
"learning_rate": 8.488146365072584e-07,
"loss": 0.4662,
"step": 442
},
{
"epoch": 0.1694072657743786,
"grad_norm": 2.7275549713665117,
"learning_rate": 8.491295484097495e-07,
"loss": 0.6759,
"step": 443
},
{
"epoch": 0.16978967495219885,
"grad_norm": 3.7458820621318125,
"learning_rate": 8.494437502509786e-07,
"loss": 0.4375,
"step": 444
},
{
"epoch": 0.1701720841300191,
"grad_norm": 4.34678554813099,
"learning_rate": 8.497572452258264e-07,
"loss": 0.3278,
"step": 445
},
{
"epoch": 0.1705544933078394,
"grad_norm": 3.6011246355051214,
"learning_rate": 8.500700365076586e-07,
"loss": 0.4652,
"step": 446
},
{
"epoch": 0.17093690248565965,
"grad_norm": 3.3932647976452524,
"learning_rate": 8.503821272485197e-07,
"loss": 0.2305,
"step": 447
},
{
"epoch": 0.1713193116634799,
"grad_norm": 3.7425874564043857,
"learning_rate": 8.506935205793227e-07,
"loss": 0.2994,
"step": 448
},
{
"epoch": 0.1717017208413002,
"grad_norm": 2.981305526120341,
"learning_rate": 8.510042196100389e-07,
"loss": 0.1297,
"step": 449
},
{
"epoch": 0.17208413001912046,
"grad_norm": 4.476331561724208,
"learning_rate": 8.513142274298821e-07,
"loss": 0.2766,
"step": 450
},
{
"epoch": 0.17246653919694072,
"grad_norm": 4.608373086952153,
"learning_rate": 8.516235471074952e-07,
"loss": 0.3614,
"step": 451
},
{
"epoch": 0.172848948374761,
"grad_norm": 4.827351233443686,
"learning_rate": 8.519321816911309e-07,
"loss": 0.5439,
"step": 452
},
{
"epoch": 0.17323135755258126,
"grad_norm": 2.511614617878112,
"learning_rate": 8.522401342088326e-07,
"loss": 0.5802,
"step": 453
},
{
"epoch": 0.17361376673040152,
"grad_norm": 3.2540966037790082,
"learning_rate": 8.525474076686125e-07,
"loss": 0.4176,
"step": 454
},
{
"epoch": 0.1739961759082218,
"grad_norm": 3.8871122994851324,
"learning_rate": 8.528540050586278e-07,
"loss": 0.6278,
"step": 455
},
{
"epoch": 0.17437858508604206,
"grad_norm": 4.048438517645086,
"learning_rate": 8.531599293473555e-07,
"loss": 0.3956,
"step": 456
},
{
"epoch": 0.17476099426386232,
"grad_norm": 3.11043051435314,
"learning_rate": 8.53465183483764e-07,
"loss": 0.2714,
"step": 457
},
{
"epoch": 0.1751434034416826,
"grad_norm": 4.921696929132415,
"learning_rate": 8.537697703974845e-07,
"loss": 0.5876,
"step": 458
},
{
"epoch": 0.17552581261950287,
"grad_norm": 4.903323521676653,
"learning_rate": 8.540736929989794e-07,
"loss": 0.4939,
"step": 459
},
{
"epoch": 0.17590822179732313,
"grad_norm": 4.8318538155200494,
"learning_rate": 8.54376954179709e-07,
"loss": 0.5429,
"step": 460
},
{
"epoch": 0.17629063097514341,
"grad_norm": 4.261696596026859,
"learning_rate": 8.546795568122967e-07,
"loss": 0.293,
"step": 461
},
{
"epoch": 0.17667304015296367,
"grad_norm": 4.632046460895395,
"learning_rate": 8.549815037506923e-07,
"loss": 0.2401,
"step": 462
},
{
"epoch": 0.17705544933078393,
"grad_norm": 8.216589582080363,
"learning_rate": 8.552827978303335e-07,
"loss": 0.6042,
"step": 463
},
{
"epoch": 0.17743785850860422,
"grad_norm": 4.647916363907478,
"learning_rate": 8.555834418683056e-07,
"loss": 0.7894,
"step": 464
},
{
"epoch": 0.17782026768642448,
"grad_norm": 2.2044464909614954,
"learning_rate": 8.558834386634994e-07,
"loss": 0.4387,
"step": 465
},
{
"epoch": 0.17820267686424474,
"grad_norm": 1.7303373058677423,
"learning_rate": 8.561827909967682e-07,
"loss": 0.4711,
"step": 466
},
{
"epoch": 0.17858508604206502,
"grad_norm": 2.3381052834893703,
"learning_rate": 8.564815016310815e-07,
"loss": 0.5498,
"step": 467
},
{
"epoch": 0.17896749521988528,
"grad_norm": 2.9010284441789858,
"learning_rate": 8.567795733116784e-07,
"loss": 0.2958,
"step": 468
},
{
"epoch": 0.17934990439770554,
"grad_norm": 2.861270463902799,
"learning_rate": 8.570770087662194e-07,
"loss": 0.3461,
"step": 469
},
{
"epoch": 0.17973231357552583,
"grad_norm": 2.3472607576215343,
"learning_rate": 8.573738107049358e-07,
"loss": 0.4536,
"step": 470
},
{
"epoch": 0.18011472275334608,
"grad_norm": 2.679119261168484,
"learning_rate": 8.576699818207777e-07,
"loss": 0.2337,
"step": 471
},
{
"epoch": 0.18049713193116634,
"grad_norm": 4.8104618264787,
"learning_rate": 8.579655247895611e-07,
"loss": 0.6163,
"step": 472
},
{
"epoch": 0.1808795411089866,
"grad_norm": 3.7416879722707956,
"learning_rate": 8.582604422701124e-07,
"loss": 0.3372,
"step": 473
},
{
"epoch": 0.1812619502868069,
"grad_norm": 3.6382333651998358,
"learning_rate": 8.585547369044125e-07,
"loss": 0.1598,
"step": 474
},
{
"epoch": 0.18164435946462715,
"grad_norm": 4.037690451650499,
"learning_rate": 8.588484113177379e-07,
"loss": 0.2873,
"step": 475
},
{
"epoch": 0.1820267686424474,
"grad_norm": 8.535887858375467,
"learning_rate": 8.591414681188027e-07,
"loss": 0.7283,
"step": 476
},
{
"epoch": 0.1824091778202677,
"grad_norm": 4.210883371999634,
"learning_rate": 8.594339098998958e-07,
"loss": 0.5389,
"step": 477
},
{
"epoch": 0.18279158699808795,
"grad_norm": 3.572301398296823,
"learning_rate": 8.597257392370203e-07,
"loss": 0.7289,
"step": 478
},
{
"epoch": 0.1831739961759082,
"grad_norm": 3.401284754868903,
"learning_rate": 8.60016958690028e-07,
"loss": 0.4667,
"step": 479
},
{
"epoch": 0.1835564053537285,
"grad_norm": 3.90239067163983,
"learning_rate": 8.603075708027559e-07,
"loss": 0.4377,
"step": 480
},
{
"epoch": 0.18393881453154876,
"grad_norm": 3.4348493586884308,
"learning_rate": 8.605975781031574e-07,
"loss": 0.4699,
"step": 481
},
{
"epoch": 0.18432122370936901,
"grad_norm": 2.926084046935673,
"learning_rate": 8.608869831034361e-07,
"loss": 0.3857,
"step": 482
},
{
"epoch": 0.1847036328871893,
"grad_norm": 3.147097621949093,
"learning_rate": 8.611757883001755e-07,
"loss": 0.2076,
"step": 483
},
{
"epoch": 0.18508604206500956,
"grad_norm": 2.9988211858901903,
"learning_rate": 8.614639961744689e-07,
"loss": 0.4364,
"step": 484
},
{
"epoch": 0.18546845124282982,
"grad_norm": 3.9578398681369062,
"learning_rate": 8.617516091920455e-07,
"loss": 0.2187,
"step": 485
},
{
"epoch": 0.1858508604206501,
"grad_norm": 2.580414743577251,
"learning_rate": 8.620386298033992e-07,
"loss": 0.1354,
"step": 486
},
{
"epoch": 0.18623326959847036,
"grad_norm": 5.465514528391503,
"learning_rate": 8.623250604439132e-07,
"loss": 0.1598,
"step": 487
},
{
"epoch": 0.18661567877629062,
"grad_norm": 5.688162628810543,
"learning_rate": 8.626109035339828e-07,
"loss": 0.5749,
"step": 488
},
{
"epoch": 0.1869980879541109,
"grad_norm": 4.06459812857032,
"learning_rate": 8.628961614791395e-07,
"loss": 0.4578,
"step": 489
},
{
"epoch": 0.18738049713193117,
"grad_norm": 4.059401753890258,
"learning_rate": 8.631808366701719e-07,
"loss": 0.5734,
"step": 490
},
{
"epoch": 0.18776290630975143,
"grad_norm": 3.6007972947017914,
"learning_rate": 8.634649314832464e-07,
"loss": 0.4735,
"step": 491
},
{
"epoch": 0.1881453154875717,
"grad_norm": 3.44509769384225,
"learning_rate": 8.637484482800254e-07,
"loss": 0.3236,
"step": 492
},
{
"epoch": 0.18852772466539197,
"grad_norm": 2.710251339310638,
"learning_rate": 8.640313894077854e-07,
"loss": 0.6083,
"step": 493
},
{
"epoch": 0.18891013384321223,
"grad_norm": 3.7640070941627126,
"learning_rate": 8.643137571995341e-07,
"loss": 0.4299,
"step": 494
},
{
"epoch": 0.18929254302103252,
"grad_norm": 11.060630511062472,
"learning_rate": 8.645955539741253e-07,
"loss": 0.2374,
"step": 495
},
{
"epoch": 0.18967495219885278,
"grad_norm": 2.9658908929801346,
"learning_rate": 8.648767820363733e-07,
"loss": 0.269,
"step": 496
},
{
"epoch": 0.19005736137667303,
"grad_norm": 4.718886215568088,
"learning_rate": 8.651574436771658e-07,
"loss": 0.3392,
"step": 497
},
{
"epoch": 0.19043977055449332,
"grad_norm": 4.07698996230184,
"learning_rate": 8.654375411735766e-07,
"loss": 0.2047,
"step": 498
},
{
"epoch": 0.19082217973231358,
"grad_norm": 4.322621506262301,
"learning_rate": 8.657170767889753e-07,
"loss": 0.1593,
"step": 499
},
{
"epoch": 0.19120458891013384,
"grad_norm": 3.6781333022238636,
"learning_rate": 8.659960527731382e-07,
"loss": 0.3298,
"step": 500
},
{
"epoch": 0.19120458891013384,
"eval_runtime": 822.3073,
"eval_samples_per_second": 1.865,
"eval_steps_per_second": 0.467,
"step": 500
},
{
"epoch": 0.19158699808795412,
"grad_norm": 5.341980758375264,
"learning_rate": 8.662744713623563e-07,
"loss": 0.5194,
"step": 501
},
{
"epoch": 0.19196940726577438,
"grad_norm": 3.8851356847921803,
"learning_rate": 8.665523347795424e-07,
"loss": 0.4366,
"step": 502
},
{
"epoch": 0.19235181644359464,
"grad_norm": 4.447133098733561,
"learning_rate": 8.668296452343392e-07,
"loss": 0.5872,
"step": 503
},
{
"epoch": 0.1927342256214149,
"grad_norm": 4.259308438470198,
"learning_rate": 8.671064049232225e-07,
"loss": 0.3353,
"step": 504
},
{
"epoch": 0.1931166347992352,
"grad_norm": 2.5481644625084,
"learning_rate": 8.673826160296078e-07,
"loss": 0.3876,
"step": 505
},
{
"epoch": 0.19349904397705545,
"grad_norm": 3.8753543814767766,
"learning_rate": 8.676582807239522e-07,
"loss": 0.6591,
"step": 506
},
{
"epoch": 0.1938814531548757,
"grad_norm": 2.9565461540677513,
"learning_rate": 8.679334011638571e-07,
"loss": 0.5829,
"step": 507
},
{
"epoch": 0.194263862332696,
"grad_norm": 2.8580497630694173,
"learning_rate": 8.682079794941694e-07,
"loss": 0.2155,
"step": 508
},
{
"epoch": 0.19464627151051625,
"grad_norm": 3.7126409535991756,
"learning_rate": 8.684820178470827e-07,
"loss": 0.3184,
"step": 509
},
{
"epoch": 0.1950286806883365,
"grad_norm": 3.961171280622527,
"learning_rate": 8.687555183422355e-07,
"loss": 0.368,
"step": 510
},
{
"epoch": 0.1954110898661568,
"grad_norm": 3.1431792391414968,
"learning_rate": 8.6902848308681e-07,
"loss": 0.1698,
"step": 511
},
{
"epoch": 0.19579349904397705,
"grad_norm": 3.4654473055387554,
"learning_rate": 8.693009141756294e-07,
"loss": 0.0964,
"step": 512
},
{
"epoch": 0.1961759082217973,
"grad_norm": 5.839838524245153,
"learning_rate": 8.695728136912551e-07,
"loss": 0.457,
"step": 513
},
{
"epoch": 0.1965583173996176,
"grad_norm": 4.7069374423354695,
"learning_rate": 8.698441837040811e-07,
"loss": 0.7124,
"step": 514
},
{
"epoch": 0.19694072657743786,
"grad_norm": 4.365320888404452,
"learning_rate": 8.701150262724292e-07,
"loss": 0.6243,
"step": 515
},
{
"epoch": 0.19732313575525812,
"grad_norm": 3.289324172386687,
"learning_rate": 8.703853434426426e-07,
"loss": 0.5249,
"step": 516
},
{
"epoch": 0.1977055449330784,
"grad_norm": 2.3283226390129723,
"learning_rate": 8.70655137249179e-07,
"loss": 0.5161,
"step": 517
},
{
"epoch": 0.19808795411089866,
"grad_norm": 2.5905756081726192,
"learning_rate": 8.709244097147014e-07,
"loss": 0.4175,
"step": 518
},
{
"epoch": 0.19847036328871892,
"grad_norm": 3.424691212788987,
"learning_rate": 8.711931628501709e-07,
"loss": 0.4771,
"step": 519
},
{
"epoch": 0.1988527724665392,
"grad_norm": 3.3766078651131006,
"learning_rate": 8.714613986549345e-07,
"loss": 0.4121,
"step": 520
},
{
"epoch": 0.19923518164435947,
"grad_norm": 3.7758399786743695,
"learning_rate": 8.717291191168164e-07,
"loss": 0.4434,
"step": 521
},
{
"epoch": 0.19961759082217972,
"grad_norm": 4.172649868787247,
"learning_rate": 8.719963262122055e-07,
"loss": 0.471,
"step": 522
},
{
"epoch": 0.2,
"grad_norm": 3.204457671034851,
"learning_rate": 8.722630219061423e-07,
"loss": 0.1419,
"step": 523
},
{
"epoch": 0.20038240917782027,
"grad_norm": 3.486618619725143,
"learning_rate": 8.725292081524071e-07,
"loss": 0.1515,
"step": 524
},
{
"epoch": 0.20076481835564053,
"grad_norm": 4.252158121924402,
"learning_rate": 8.72794886893605e-07,
"loss": 0.2632,
"step": 525
},
{
"epoch": 0.20114722753346082,
"grad_norm": 8.937775720998328,
"learning_rate": 8.730600600612507e-07,
"loss": 0.4893,
"step": 526
},
{
"epoch": 0.20152963671128107,
"grad_norm": 4.2345114823195935,
"learning_rate": 8.73324729575853e-07,
"loss": 0.527,
"step": 527
},
{
"epoch": 0.20191204588910133,
"grad_norm": 3.6419135829805787,
"learning_rate": 8.73588897346999e-07,
"loss": 0.6547,
"step": 528
},
{
"epoch": 0.20229445506692162,
"grad_norm": 1.8780114468031686,
"learning_rate": 8.738525652734356e-07,
"loss": 0.563,
"step": 529
},
{
"epoch": 0.20267686424474188,
"grad_norm": 3.003413959935241,
"learning_rate": 8.74115735243152e-07,
"loss": 0.6466,
"step": 530
},
{
"epoch": 0.20305927342256214,
"grad_norm": 3.2604255716773025,
"learning_rate": 8.743784091334607e-07,
"loss": 0.4341,
"step": 531
},
{
"epoch": 0.20344168260038242,
"grad_norm": 2.983473237094229,
"learning_rate": 8.746405888110783e-07,
"loss": 0.3732,
"step": 532
},
{
"epoch": 0.20382409177820268,
"grad_norm": 4.1953099994900604,
"learning_rate": 8.74902276132204e-07,
"loss": 0.4144,
"step": 533
},
{
"epoch": 0.20420650095602294,
"grad_norm": 3.45385513526699,
"learning_rate": 8.751634729425997e-07,
"loss": 0.1678,
"step": 534
},
{
"epoch": 0.2045889101338432,
"grad_norm": 3.767437257402485,
"learning_rate": 8.754241810776672e-07,
"loss": 0.2739,
"step": 535
},
{
"epoch": 0.20497131931166349,
"grad_norm": 4.0418594681414985,
"learning_rate": 8.756844023625261e-07,
"loss": 0.1897,
"step": 536
},
{
"epoch": 0.20535372848948374,
"grad_norm": 4.170464094943381,
"learning_rate": 8.7594413861209e-07,
"loss": 0.0853,
"step": 537
},
{
"epoch": 0.205736137667304,
"grad_norm": 7.11126053507429,
"learning_rate": 8.762033916311432e-07,
"loss": 0.2841,
"step": 538
},
{
"epoch": 0.2061185468451243,
"grad_norm": 8.038682400361811,
"learning_rate": 8.764621632144152e-07,
"loss": 0.5757,
"step": 539
},
{
"epoch": 0.20650095602294455,
"grad_norm": 3.8845616167680714,
"learning_rate": 8.767204551466555e-07,
"loss": 0.5021,
"step": 540
},
{
"epoch": 0.2068833652007648,
"grad_norm": 3.457770765270759,
"learning_rate": 8.769782692027076e-07,
"loss": 0.4517,
"step": 541
},
{
"epoch": 0.2072657743785851,
"grad_norm": 2.1231287402641406,
"learning_rate": 8.77235607147582e-07,
"loss": 0.406,
"step": 542
},
{
"epoch": 0.20764818355640535,
"grad_norm": 3.664290236042986,
"learning_rate": 8.774924707365287e-07,
"loss": 0.384,
"step": 543
},
{
"epoch": 0.2080305927342256,
"grad_norm": 3.8846193097131776,
"learning_rate": 8.777488617151092e-07,
"loss": 0.5264,
"step": 544
},
{
"epoch": 0.2084130019120459,
"grad_norm": 3.8864727613347774,
"learning_rate": 8.780047818192677e-07,
"loss": 0.3449,
"step": 545
},
{
"epoch": 0.20879541108986616,
"grad_norm": 3.074034475695364,
"learning_rate": 8.782602327754011e-07,
"loss": 0.4267,
"step": 546
},
{
"epoch": 0.20917782026768642,
"grad_norm": 4.650051420194106,
"learning_rate": 8.785152163004298e-07,
"loss": 0.2878,
"step": 547
},
{
"epoch": 0.2095602294455067,
"grad_norm": 3.9610133417757614,
"learning_rate": 8.787697341018658e-07,
"loss": 0.2085,
"step": 548
},
{
"epoch": 0.20994263862332696,
"grad_norm": 4.055132009142522,
"learning_rate": 8.790237878778824e-07,
"loss": 0.1718,
"step": 549
},
{
"epoch": 0.21032504780114722,
"grad_norm": 5.561179357908723,
"learning_rate": 8.792773793173814e-07,
"loss": 0.2292,
"step": 550
},
{
"epoch": 0.2107074569789675,
"grad_norm": 9.823343493565627,
"learning_rate": 8.795305101000612e-07,
"loss": 0.4747,
"step": 551
},
{
"epoch": 0.21108986615678776,
"grad_norm": 4.320240537516627,
"learning_rate": 8.797831818964823e-07,
"loss": 0.6607,
"step": 552
},
{
"epoch": 0.21147227533460802,
"grad_norm": 3.8821286495378406,
"learning_rate": 8.800353963681352e-07,
"loss": 0.6069,
"step": 553
},
{
"epoch": 0.2118546845124283,
"grad_norm": 3.456264535828526,
"learning_rate": 8.802871551675046e-07,
"loss": 0.5072,
"step": 554
},
{
"epoch": 0.21223709369024857,
"grad_norm": 4.004620506032858,
"learning_rate": 8.805384599381345e-07,
"loss": 0.3927,
"step": 555
},
{
"epoch": 0.21261950286806883,
"grad_norm": 3.7279097642824355,
"learning_rate": 8.807893123146934e-07,
"loss": 0.5367,
"step": 556
},
{
"epoch": 0.2130019120458891,
"grad_norm": 2.8182120069861316,
"learning_rate": 8.810397139230372e-07,
"loss": 0.4391,
"step": 557
},
{
"epoch": 0.21338432122370937,
"grad_norm": 2.7487648877915536,
"learning_rate": 8.812896663802729e-07,
"loss": 0.2057,
"step": 558
},
{
"epoch": 0.21376673040152963,
"grad_norm": 4.265964592351681,
"learning_rate": 8.815391712948212e-07,
"loss": 0.3774,
"step": 559
},
{
"epoch": 0.21414913957934992,
"grad_norm": 4.349966307761718,
"learning_rate": 8.817882302664786e-07,
"loss": 0.2399,
"step": 560
},
{
"epoch": 0.21453154875717018,
"grad_norm": 4.632009813011725,
"learning_rate": 8.820368448864787e-07,
"loss": 0.1874,
"step": 561
},
{
"epoch": 0.21491395793499043,
"grad_norm": 5.107100700943744,
"learning_rate": 8.822850167375535e-07,
"loss": 0.1995,
"step": 562
},
{
"epoch": 0.2152963671128107,
"grad_norm": 9.613688373896391,
"learning_rate": 8.82532747393994e-07,
"loss": 0.5337,
"step": 563
},
{
"epoch": 0.21567877629063098,
"grad_norm": 4.014300294858346,
"learning_rate": 8.827800384217091e-07,
"loss": 0.4894,
"step": 564
},
{
"epoch": 0.21606118546845124,
"grad_norm": 3.7285606417528623,
"learning_rate": 8.830268913782868e-07,
"loss": 0.564,
"step": 565
},
{
"epoch": 0.2164435946462715,
"grad_norm": 4.170282066923075,
"learning_rate": 8.832733078130512e-07,
"loss": 0.7278,
"step": 566
},
{
"epoch": 0.21682600382409178,
"grad_norm": 2.822889339600972,
"learning_rate": 8.835192892671222e-07,
"loss": 0.5958,
"step": 567
},
{
"epoch": 0.21720841300191204,
"grad_norm": 2.933753526097939,
"learning_rate": 8.837648372734725e-07,
"loss": 0.3823,
"step": 568
},
{
"epoch": 0.2175908221797323,
"grad_norm": 3.8063780614296845,
"learning_rate": 8.840099533569853e-07,
"loss": 0.4028,
"step": 569
},
{
"epoch": 0.2179732313575526,
"grad_norm": 5.715698094888678,
"learning_rate": 8.842546390345113e-07,
"loss": 0.4984,
"step": 570
},
{
"epoch": 0.21835564053537285,
"grad_norm": 3.3179041127519064,
"learning_rate": 8.844988958149243e-07,
"loss": 0.4208,
"step": 571
},
{
"epoch": 0.2187380497131931,
"grad_norm": 3.13872732648784,
"learning_rate": 8.847427251991777e-07,
"loss": 0.2319,
"step": 572
},
{
"epoch": 0.2191204588910134,
"grad_norm": 2.9904045191122965,
"learning_rate": 8.849861286803598e-07,
"loss": 0.3062,
"step": 573
},
{
"epoch": 0.21950286806883365,
"grad_norm": 4.134629771253638,
"learning_rate": 8.852291077437482e-07,
"loss": 0.1593,
"step": 574
},
{
"epoch": 0.2198852772466539,
"grad_norm": 5.5330836818095595,
"learning_rate": 8.854716638668649e-07,
"loss": 0.3453,
"step": 575
},
{
"epoch": 0.2202676864244742,
"grad_norm": 6.35525536721151,
"learning_rate": 8.857137985195292e-07,
"loss": 0.3971,
"step": 576
},
{
"epoch": 0.22065009560229445,
"grad_norm": 3.7363729200055866,
"learning_rate": 8.859555131639122e-07,
"loss": 0.4838,
"step": 577
},
{
"epoch": 0.2210325047801147,
"grad_norm": 2.884765790294934,
"learning_rate": 8.86196809254589e-07,
"loss": 0.3447,
"step": 578
},
{
"epoch": 0.221414913957935,
"grad_norm": 2.6882944041880656,
"learning_rate": 8.864376882385918e-07,
"loss": 0.6815,
"step": 579
},
{
"epoch": 0.22179732313575526,
"grad_norm": 2.6886467372301643,
"learning_rate": 8.866781515554614e-07,
"loss": 0.3703,
"step": 580
},
{
"epoch": 0.22217973231357552,
"grad_norm": 2.136208203008963,
"learning_rate": 8.869182006372994e-07,
"loss": 0.4558,
"step": 581
},
{
"epoch": 0.2225621414913958,
"grad_norm": 2.9403023845844443,
"learning_rate": 8.871578369088188e-07,
"loss": 0.2963,
"step": 582
},
{
"epoch": 0.22294455066921606,
"grad_norm": 2.875306486296151,
"learning_rate": 8.873970617873952e-07,
"loss": 0.4421,
"step": 583
},
{
"epoch": 0.22332695984703632,
"grad_norm": 4.129625905530537,
"learning_rate": 8.876358766831167e-07,
"loss": 0.2216,
"step": 584
},
{
"epoch": 0.2237093690248566,
"grad_norm": 3.3319302148918393,
"learning_rate": 8.878742829988343e-07,
"loss": 0.2468,
"step": 585
},
{
"epoch": 0.22409177820267687,
"grad_norm": 4.4162868600360365,
"learning_rate": 8.881122821302105e-07,
"loss": 0.1727,
"step": 586
},
{
"epoch": 0.22447418738049713,
"grad_norm": 3.5349965433538175,
"learning_rate": 8.883498754657699e-07,
"loss": 0.2095,
"step": 587
},
{
"epoch": 0.2248565965583174,
"grad_norm": 7.335889399719171,
"learning_rate": 8.885870643869454e-07,
"loss": 0.5801,
"step": 588
},
{
"epoch": 0.22523900573613767,
"grad_norm": 3.8911290589093004,
"learning_rate": 8.888238502681287e-07,
"loss": 0.4556,
"step": 589
},
{
"epoch": 0.22562141491395793,
"grad_norm": 2.9425051638704556,
"learning_rate": 8.890602344767169e-07,
"loss": 0.5179,
"step": 590
},
{
"epoch": 0.22600382409177822,
"grad_norm": 2.409984514495357,
"learning_rate": 8.892962183731602e-07,
"loss": 0.821,
"step": 591
},
{
"epoch": 0.22638623326959847,
"grad_norm": 4.32976382521085,
"learning_rate": 8.895318033110082e-07,
"loss": 0.6991,
"step": 592
},
{
"epoch": 0.22676864244741873,
"grad_norm": 6.219767271637266,
"learning_rate": 8.897669906369579e-07,
"loss": 0.5052,
"step": 593
},
{
"epoch": 0.227151051625239,
"grad_norm": 2.847185009555525,
"learning_rate": 8.900017816908987e-07,
"loss": 0.3373,
"step": 594
},
{
"epoch": 0.22753346080305928,
"grad_norm": 2.5647512641752654,
"learning_rate": 8.902361778059585e-07,
"loss": 0.3058,
"step": 595
},
{
"epoch": 0.22791586998087954,
"grad_norm": 3.2259829834349127,
"learning_rate": 8.904701803085492e-07,
"loss": 0.4217,
"step": 596
},
{
"epoch": 0.2282982791586998,
"grad_norm": 3.240345774202062,
"learning_rate": 8.907037905184118e-07,
"loss": 0.288,
"step": 597
},
{
"epoch": 0.22868068833652008,
"grad_norm": 2.8117471323067367,
"learning_rate": 8.90937009748661e-07,
"loss": 0.3434,
"step": 598
},
{
"epoch": 0.22906309751434034,
"grad_norm": 3.409382351064056,
"learning_rate": 8.911698393058294e-07,
"loss": 0.1568,
"step": 599
},
{
"epoch": 0.2294455066921606,
"grad_norm": 5.025906586355392,
"learning_rate": 8.914022804899116e-07,
"loss": 0.2689,
"step": 600
},
{
"epoch": 0.2298279158699809,
"grad_norm": 5.246780134732783,
"learning_rate": 8.916343345944075e-07,
"loss": 0.5604,
"step": 601
},
{
"epoch": 0.23021032504780115,
"grad_norm": 4.418125895193714,
"learning_rate": 8.918660029063655e-07,
"loss": 0.6224,
"step": 602
},
{
"epoch": 0.2305927342256214,
"grad_norm": 10.199199301015543,
"learning_rate": 8.920972867064258e-07,
"loss": 0.3815,
"step": 603
},
{
"epoch": 0.2309751434034417,
"grad_norm": 2.503913486607132,
"learning_rate": 8.923281872688621e-07,
"loss": 0.241,
"step": 604
},
{
"epoch": 0.23135755258126195,
"grad_norm": 2.652775185823977,
"learning_rate": 8.925587058616247e-07,
"loss": 0.4143,
"step": 605
},
{
"epoch": 0.2317399617590822,
"grad_norm": 2.9848565317305704,
"learning_rate": 8.927888437463812e-07,
"loss": 0.6514,
"step": 606
},
{
"epoch": 0.2321223709369025,
"grad_norm": 4.13808400622241,
"learning_rate": 8.930186021785589e-07,
"loss": 0.5333,
"step": 607
},
{
"epoch": 0.23250478011472275,
"grad_norm": 2.6761707774182173,
"learning_rate": 8.93247982407385e-07,
"loss": 0.4142,
"step": 608
},
{
"epoch": 0.232887189292543,
"grad_norm": 4.37456401287424,
"learning_rate": 8.934769856759282e-07,
"loss": 0.3991,
"step": 609
},
{
"epoch": 0.2332695984703633,
"grad_norm": 4.247755916759948,
"learning_rate": 8.937056132211385e-07,
"loss": 0.299,
"step": 610
},
{
"epoch": 0.23365200764818356,
"grad_norm": 2.9509372580052586,
"learning_rate": 8.939338662738878e-07,
"loss": 0.1705,
"step": 611
},
{
"epoch": 0.23403441682600382,
"grad_norm": 4.224394914083757,
"learning_rate": 8.94161746059009e-07,
"loss": 0.2091,
"step": 612
},
{
"epoch": 0.2344168260038241,
"grad_norm": 5.198523150312591,
"learning_rate": 8.943892537953359e-07,
"loss": 0.4931,
"step": 613
},
{
"epoch": 0.23479923518164436,
"grad_norm": 5.9789195081642825,
"learning_rate": 8.946163906957424e-07,
"loss": 0.4767,
"step": 614
},
{
"epoch": 0.23518164435946462,
"grad_norm": 3.501394266617619,
"learning_rate": 8.948431579671812e-07,
"loss": 0.7518,
"step": 615
},
{
"epoch": 0.2355640535372849,
"grad_norm": 2.534678080506824,
"learning_rate": 8.950695568107218e-07,
"loss": 0.6598,
"step": 616
},
{
"epoch": 0.23594646271510517,
"grad_norm": 3.2492748331786774,
"learning_rate": 8.952955884215896e-07,
"loss": 0.6163,
"step": 617
},
{
"epoch": 0.23632887189292542,
"grad_norm": 2.9826559338411354,
"learning_rate": 8.955212539892024e-07,
"loss": 0.3615,
"step": 618
},
{
"epoch": 0.2367112810707457,
"grad_norm": 3.318038053711697,
"learning_rate": 8.957465546972099e-07,
"loss": 0.5365,
"step": 619
},
{
"epoch": 0.23709369024856597,
"grad_norm": 3.4175041536921253,
"learning_rate": 8.959714917235292e-07,
"loss": 0.2957,
"step": 620
},
{
"epoch": 0.23747609942638623,
"grad_norm": 4.206538406165909,
"learning_rate": 8.96196066240382e-07,
"loss": 0.5224,
"step": 621
},
{
"epoch": 0.23785850860420651,
"grad_norm": 3.2186359274196215,
"learning_rate": 8.964202794143326e-07,
"loss": 0.1611,
"step": 622
},
{
"epoch": 0.23824091778202677,
"grad_norm": 3.1131388058497014,
"learning_rate": 8.966441324063227e-07,
"loss": 0.3956,
"step": 623
},
{
"epoch": 0.23862332695984703,
"grad_norm": 2.889521867739628,
"learning_rate": 8.968676263717078e-07,
"loss": 0.172,
"step": 624
},
{
"epoch": 0.2390057361376673,
"grad_norm": 5.782986436859959,
"learning_rate": 8.970907624602941e-07,
"loss": 0.183,
"step": 625
},
{
"epoch": 0.23938814531548758,
"grad_norm": 8.177253193316057,
"learning_rate": 8.973135418163724e-07,
"loss": 0.4219,
"step": 626
},
{
"epoch": 0.23977055449330784,
"grad_norm": 3.4656521822772333,
"learning_rate": 8.975359655787545e-07,
"loss": 0.6771,
"step": 627
},
{
"epoch": 0.2401529636711281,
"grad_norm": 3.9809862274031342,
"learning_rate": 8.977580348808072e-07,
"loss": 0.4621,
"step": 628
},
{
"epoch": 0.24053537284894838,
"grad_norm": 2.590580301854008,
"learning_rate": 8.97979750850488e-07,
"loss": 0.3399,
"step": 629
},
{
"epoch": 0.24091778202676864,
"grad_norm": 2.797762426317898,
"learning_rate": 8.982011146103783e-07,
"loss": 0.2546,
"step": 630
},
{
"epoch": 0.2413001912045889,
"grad_norm": 3.0114940752468007,
"learning_rate": 8.984221272777187e-07,
"loss": 0.1752,
"step": 631
},
{
"epoch": 0.24168260038240919,
"grad_norm": 3.0610270169202085,
"learning_rate": 8.986427899644419e-07,
"loss": 0.3983,
"step": 632
},
{
"epoch": 0.24206500956022944,
"grad_norm": 2.8067468080231506,
"learning_rate": 8.988631037772066e-07,
"loss": 0.2748,
"step": 633
},
{
"epoch": 0.2424474187380497,
"grad_norm": 4.000175233817278,
"learning_rate": 8.990830698174309e-07,
"loss": 0.1971,
"step": 634
},
{
"epoch": 0.24282982791587,
"grad_norm": 3.769745657560052,
"learning_rate": 8.993026891813253e-07,
"loss": 0.4406,
"step": 635
},
{
"epoch": 0.24321223709369025,
"grad_norm": 4.197854974765107,
"learning_rate": 8.995219629599254e-07,
"loss": 0.243,
"step": 636
},
{
"epoch": 0.2435946462715105,
"grad_norm": 2.9496964831922186,
"learning_rate": 8.997408922391241e-07,
"loss": 0.1551,
"step": 637
},
{
"epoch": 0.2439770554493308,
"grad_norm": 7.136065540371976,
"learning_rate": 8.999594780997046e-07,
"loss": 0.5281,
"step": 638
},
{
"epoch": 0.24435946462715105,
"grad_norm": 3.1328861675020034,
"learning_rate": 9.001777216173722e-07,
"loss": 0.4891,
"step": 639
},
{
"epoch": 0.2447418738049713,
"grad_norm": 3.851332837544597,
"learning_rate": 9.003956238627853e-07,
"loss": 0.478,
"step": 640
},
{
"epoch": 0.2451242829827916,
"grad_norm": 2.725577365010256,
"learning_rate": 9.00613185901588e-07,
"loss": 0.4471,
"step": 641
},
{
"epoch": 0.24550669216061186,
"grad_norm": 2.5746230877323146,
"learning_rate": 9.008304087944406e-07,
"loss": 0.3046,
"step": 642
},
{
"epoch": 0.24588910133843211,
"grad_norm": 4.590008308220286,
"learning_rate": 9.010472935970509e-07,
"loss": 0.343,
"step": 643
},
{
"epoch": 0.2462715105162524,
"grad_norm": 3.231391156948366,
"learning_rate": 9.012638413602053e-07,
"loss": 0.3504,
"step": 644
},
{
"epoch": 0.24665391969407266,
"grad_norm": 2.5125683523720896,
"learning_rate": 9.014800531297985e-07,
"loss": 0.2493,
"step": 645
},
{
"epoch": 0.24703632887189292,
"grad_norm": 2.79504368674542,
"learning_rate": 9.016959299468647e-07,
"loss": 0.2146,
"step": 646
},
{
"epoch": 0.2474187380497132,
"grad_norm": 3.8159029422788753,
"learning_rate": 9.019114728476074e-07,
"loss": 0.4076,
"step": 647
},
{
"epoch": 0.24780114722753346,
"grad_norm": 4.783488030319095,
"learning_rate": 9.021266828634289e-07,
"loss": 0.1647,
"step": 648
},
{
"epoch": 0.24818355640535372,
"grad_norm": 6.686053253263,
"learning_rate": 9.023415610209602e-07,
"loss": 0.1261,
"step": 649
},
{
"epoch": 0.248565965583174,
"grad_norm": 4.911620098644459,
"learning_rate": 9.025561083420904e-07,
"loss": 0.3726,
"step": 650
},
{
"epoch": 0.24894837476099427,
"grad_norm": 6.45514667607322,
"learning_rate": 9.027703258439958e-07,
"loss": 0.6213,
"step": 651
},
{
"epoch": 0.24933078393881453,
"grad_norm": 3.249407494161105,
"learning_rate": 9.029842145391691e-07,
"loss": 0.6847,
"step": 652
},
{
"epoch": 0.2497131931166348,
"grad_norm": 4.2437066998751725,
"learning_rate": 9.031977754354472e-07,
"loss": 0.5215,
"step": 653
},
{
"epoch": 0.25009560229445504,
"grad_norm": 1.7989403837715172,
"learning_rate": 9.034110095360411e-07,
"loss": 0.543,
"step": 654
},
{
"epoch": 0.25047801147227533,
"grad_norm": 2.760700388788634,
"learning_rate": 9.03623917839563e-07,
"loss": 0.3404,
"step": 655
},
{
"epoch": 0.2508604206500956,
"grad_norm": 3.1453054292526397,
"learning_rate": 9.038365013400548e-07,
"loss": 0.4307,
"step": 656
},
{
"epoch": 0.25124282982791585,
"grad_norm": 2.750521350338774,
"learning_rate": 9.040487610270163e-07,
"loss": 0.3821,
"step": 657
},
{
"epoch": 0.25162523900573613,
"grad_norm": 4.3433778452452625,
"learning_rate": 9.04260697885432e-07,
"loss": 0.4657,
"step": 658
},
{
"epoch": 0.2520076481835564,
"grad_norm": 3.038734869470121,
"learning_rate": 9.044723128957992e-07,
"loss": 0.4424,
"step": 659
},
{
"epoch": 0.25239005736137665,
"grad_norm": 3.6418428798847953,
"learning_rate": 9.046836070341548e-07,
"loss": 0.1612,
"step": 660
},
{
"epoch": 0.25277246653919694,
"grad_norm": 4.814089029716334,
"learning_rate": 9.04894581272103e-07,
"loss": 0.1493,
"step": 661
},
{
"epoch": 0.2531548757170172,
"grad_norm": 4.9591074534253785,
"learning_rate": 9.051052365768415e-07,
"loss": 0.1937,
"step": 662
},
{
"epoch": 0.25353728489483746,
"grad_norm": 3.944316957470118,
"learning_rate": 9.053155739111876e-07,
"loss": 0.3188,
"step": 663
},
{
"epoch": 0.25391969407265774,
"grad_norm": 4.223668375030919,
"learning_rate": 9.055255942336061e-07,
"loss": 0.6315,
"step": 664
},
{
"epoch": 0.25430210325047803,
"grad_norm": 3.4732364817739083,
"learning_rate": 9.057352984982342e-07,
"loss": 0.6295,
"step": 665
},
{
"epoch": 0.25468451242829826,
"grad_norm": 3.165842718713124,
"learning_rate": 9.05944687654908e-07,
"loss": 0.689,
"step": 666
},
{
"epoch": 0.25506692160611855,
"grad_norm": 2.3999509726920847,
"learning_rate": 9.06153762649188e-07,
"loss": 0.3787,
"step": 667
},
{
"epoch": 0.25544933078393883,
"grad_norm": 3.3319283544432743,
"learning_rate": 9.063625244223857e-07,
"loss": 0.4465,
"step": 668
},
{
"epoch": 0.25583173996175906,
"grad_norm": 2.5320497356188447,
"learning_rate": 9.065709739115878e-07,
"loss": 0.4041,
"step": 669
},
{
"epoch": 0.25621414913957935,
"grad_norm": 3.306812513990708,
"learning_rate": 9.067791120496818e-07,
"loss": 0.1911,
"step": 670
},
{
"epoch": 0.25659655831739964,
"grad_norm": 4.444888988091275,
"learning_rate": 9.069869397653817e-07,
"loss": 0.1971,
"step": 671
},
{
"epoch": 0.25697896749521987,
"grad_norm": 2.9716506440106363,
"learning_rate": 9.071944579832521e-07,
"loss": 0.1668,
"step": 672
},
{
"epoch": 0.25736137667304015,
"grad_norm": 4.030938661366508,
"learning_rate": 9.074016676237329e-07,
"loss": 0.2018,
"step": 673
},
{
"epoch": 0.25774378585086044,
"grad_norm": 3.297851128729243,
"learning_rate": 9.076085696031645e-07,
"loss": 0.1608,
"step": 674
},
{
"epoch": 0.25812619502868067,
"grad_norm": 4.587013339388978,
"learning_rate": 9.078151648338114e-07,
"loss": 0.3073,
"step": 675
},
{
"epoch": 0.25850860420650096,
"grad_norm": 5.998579174931337,
"learning_rate": 9.080214542238866e-07,
"loss": 0.3133,
"step": 676
},
{
"epoch": 0.25889101338432124,
"grad_norm": 3.426778824253407,
"learning_rate": 9.082274386775756e-07,
"loss": 0.6327,
"step": 677
},
{
"epoch": 0.2592734225621415,
"grad_norm": 2.835126944900558,
"learning_rate": 9.0843311909506e-07,
"loss": 0.7641,
"step": 678
},
{
"epoch": 0.25965583173996176,
"grad_norm": 3.2794105823608297,
"learning_rate": 9.086384963725418e-07,
"loss": 0.359,
"step": 679
},
{
"epoch": 0.26003824091778205,
"grad_norm": 2.7284902845883954,
"learning_rate": 9.088435714022651e-07,
"loss": 0.3456,
"step": 680
},
{
"epoch": 0.2604206500956023,
"grad_norm": 3.007036698599815,
"learning_rate": 9.090483450725417e-07,
"loss": 0.2821,
"step": 681
},
{
"epoch": 0.26080305927342257,
"grad_norm": 2.9265717481313267,
"learning_rate": 9.092528182677723e-07,
"loss": 0.5039,
"step": 682
},
{
"epoch": 0.26118546845124285,
"grad_norm": 2.9918622270880832,
"learning_rate": 9.094569918684705e-07,
"loss": 0.3882,
"step": 683
},
{
"epoch": 0.2615678776290631,
"grad_norm": 2.9372459464087823,
"learning_rate": 9.096608667512846e-07,
"loss": 0.2574,
"step": 684
},
{
"epoch": 0.26195028680688337,
"grad_norm": 2.7236732542824864,
"learning_rate": 9.098644437890217e-07,
"loss": 0.1896,
"step": 685
},
{
"epoch": 0.26233269598470366,
"grad_norm": 4.740230753629643,
"learning_rate": 9.100677238506682e-07,
"loss": 0.181,
"step": 686
},
{
"epoch": 0.2627151051625239,
"grad_norm": 4.5618603401664535,
"learning_rate": 9.102707078014137e-07,
"loss": 0.2754,
"step": 687
},
{
"epoch": 0.2630975143403442,
"grad_norm": 5.453959841069427,
"learning_rate": 9.104733965026722e-07,
"loss": 0.4325,
"step": 688
},
{
"epoch": 0.26347992351816446,
"grad_norm": 6.570727903318121,
"learning_rate": 9.106757908121041e-07,
"loss": 0.6136,
"step": 689
},
{
"epoch": 0.2638623326959847,
"grad_norm": 3.123896136824884,
"learning_rate": 9.108778915836382e-07,
"loss": 0.3504,
"step": 690
},
{
"epoch": 0.264244741873805,
"grad_norm": 3.4093210917126133,
"learning_rate": 9.110796996674936e-07,
"loss": 0.6091,
"step": 691
},
{
"epoch": 0.26462715105162526,
"grad_norm": 4.880015409789222,
"learning_rate": 9.112812159102003e-07,
"loss": 0.5309,
"step": 692
},
{
"epoch": 0.2650095602294455,
"grad_norm": 3.3460021233332453,
"learning_rate": 9.114824411546215e-07,
"loss": 0.4077,
"step": 693
},
{
"epoch": 0.2653919694072658,
"grad_norm": 2.398058833446472,
"learning_rate": 9.116833762399739e-07,
"loss": 0.3564,
"step": 694
},
{
"epoch": 0.26577437858508607,
"grad_norm": 3.449133037050817,
"learning_rate": 9.118840220018492e-07,
"loss": 0.1793,
"step": 695
},
{
"epoch": 0.2661567877629063,
"grad_norm": 3.9880952109829617,
"learning_rate": 9.120843792722349e-07,
"loss": 0.3157,
"step": 696
},
{
"epoch": 0.2665391969407266,
"grad_norm": 3.666445869964919,
"learning_rate": 9.122844488795347e-07,
"loss": 0.2338,
"step": 697
},
{
"epoch": 0.2669216061185469,
"grad_norm": 4.149391883508918,
"learning_rate": 9.124842316485896e-07,
"loss": 0.1581,
"step": 698
},
{
"epoch": 0.2673040152963671,
"grad_norm": 3.5721827310212104,
"learning_rate": 9.126837284006975e-07,
"loss": 0.1278,
"step": 699
},
{
"epoch": 0.2676864244741874,
"grad_norm": 4.706442844447054,
"learning_rate": 9.128829399536345e-07,
"loss": 0.2241,
"step": 700
},
{
"epoch": 0.2680688336520076,
"grad_norm": 10.142777271838952,
"learning_rate": 9.130818671216741e-07,
"loss": 0.5104,
"step": 701
},
{
"epoch": 0.2684512428298279,
"grad_norm": 4.933768492055857,
"learning_rate": 9.132805107156076e-07,
"loss": 0.5941,
"step": 702
},
{
"epoch": 0.2688336520076482,
"grad_norm": 2.7437742158346863,
"learning_rate": 9.134788715427637e-07,
"loss": 0.2277,
"step": 703
},
{
"epoch": 0.2692160611854684,
"grad_norm": 3.2670954993715995,
"learning_rate": 9.136769504070285e-07,
"loss": 0.5396,
"step": 704
},
{
"epoch": 0.2695984703632887,
"grad_norm": 2.0297961387309633,
"learning_rate": 9.13874748108865e-07,
"loss": 0.4631,
"step": 705
},
{
"epoch": 0.269980879541109,
"grad_norm": 6.472273452112172,
"learning_rate": 9.14072265445332e-07,
"loss": 0.481,
"step": 706
},
{
"epoch": 0.27036328871892923,
"grad_norm": 3.651355173071301,
"learning_rate": 9.142695032101042e-07,
"loss": 0.4247,
"step": 707
},
{
"epoch": 0.2707456978967495,
"grad_norm": 3.490073569104415,
"learning_rate": 9.144664621934904e-07,
"loss": 0.2685,
"step": 708
},
{
"epoch": 0.2711281070745698,
"grad_norm": 3.3061256255497162,
"learning_rate": 9.146631431824534e-07,
"loss": 0.4662,
"step": 709
},
{
"epoch": 0.27151051625239003,
"grad_norm": 4.023618112602692,
"learning_rate": 9.148595469606283e-07,
"loss": 0.243,
"step": 710
},
{
"epoch": 0.2718929254302103,
"grad_norm": 4.451440110939301,
"learning_rate": 9.150556743083417e-07,
"loss": 0.4252,
"step": 711
},
{
"epoch": 0.2722753346080306,
"grad_norm": 3.282564898834976,
"learning_rate": 9.152515260026294e-07,
"loss": 0.1568,
"step": 712
},
{
"epoch": 0.27265774378585084,
"grad_norm": 5.136206678967686,
"learning_rate": 9.154471028172556e-07,
"loss": 0.3462,
"step": 713
},
{
"epoch": 0.2730401529636711,
"grad_norm": 3.765053860100544,
"learning_rate": 9.156424055227317e-07,
"loss": 0.6971,
"step": 714
},
{
"epoch": 0.2734225621414914,
"grad_norm": 4.7913020561927375,
"learning_rate": 9.158374348863336e-07,
"loss": 0.6216,
"step": 715
},
{
"epoch": 0.27380497131931164,
"grad_norm": 2.896510879693719,
"learning_rate": 9.160321916721196e-07,
"loss": 0.5049,
"step": 716
},
{
"epoch": 0.2741873804971319,
"grad_norm": 3.0359787385017647,
"learning_rate": 9.162266766409494e-07,
"loss": 0.4884,
"step": 717
},
{
"epoch": 0.2745697896749522,
"grad_norm": 4.12477504232416,
"learning_rate": 9.164208905505015e-07,
"loss": 0.3276,
"step": 718
},
{
"epoch": 0.27495219885277244,
"grad_norm": 2.762041690834952,
"learning_rate": 9.166148341552905e-07,
"loss": 0.3178,
"step": 719
},
{
"epoch": 0.27533460803059273,
"grad_norm": 2.7617872949966977,
"learning_rate": 9.16808508206685e-07,
"loss": 0.2372,
"step": 720
},
{
"epoch": 0.275717017208413,
"grad_norm": 6.980629140487373,
"learning_rate": 9.170019134529255e-07,
"loss": 0.5347,
"step": 721
},
{
"epoch": 0.27609942638623325,
"grad_norm": 4.8218599923692045,
"learning_rate": 9.171950506391405e-07,
"loss": 0.3513,
"step": 722
},
{
"epoch": 0.27648183556405354,
"grad_norm": 4.577506274183089,
"learning_rate": 9.173879205073653e-07,
"loss": 0.1856,
"step": 723
},
{
"epoch": 0.2768642447418738,
"grad_norm": 3.8802146280054157,
"learning_rate": 9.175805237965581e-07,
"loss": 0.1468,
"step": 724
},
{
"epoch": 0.27724665391969405,
"grad_norm": 5.233321627626063,
"learning_rate": 9.177728612426173e-07,
"loss": 0.183,
"step": 725
},
{
"epoch": 0.27762906309751434,
"grad_norm": 4.871894434118751,
"learning_rate": 9.17964933578398e-07,
"loss": 0.5844,
"step": 726
},
{
"epoch": 0.2780114722753346,
"grad_norm": 3.7483749683375165,
"learning_rate": 9.181567415337294e-07,
"loss": 0.3946,
"step": 727
},
{
"epoch": 0.27839388145315486,
"grad_norm": 3.927752221141394,
"learning_rate": 9.183482858354308e-07,
"loss": 0.7132,
"step": 728
},
{
"epoch": 0.27877629063097514,
"grad_norm": 2.509795833658178,
"learning_rate": 9.185395672073286e-07,
"loss": 0.4222,
"step": 729
},
{
"epoch": 0.27915869980879543,
"grad_norm": 3.4221312544513647,
"learning_rate": 9.187305863702726e-07,
"loss": 0.501,
"step": 730
},
{
"epoch": 0.27954110898661566,
"grad_norm": 3.2600125355859806,
"learning_rate": 9.189213440421519e-07,
"loss": 0.4319,
"step": 731
},
{
"epoch": 0.27992351816443595,
"grad_norm": 3.6358730754710153,
"learning_rate": 9.191118409379119e-07,
"loss": 0.5501,
"step": 732
},
{
"epoch": 0.28030592734225623,
"grad_norm": 2.838876425692313,
"learning_rate": 9.193020777695696e-07,
"loss": 0.4226,
"step": 733
},
{
"epoch": 0.28068833652007646,
"grad_norm": 4.23248541643004,
"learning_rate": 9.194920552462298e-07,
"loss": 0.2216,
"step": 734
},
{
"epoch": 0.28107074569789675,
"grad_norm": 5.05975286098859,
"learning_rate": 9.196817740741012e-07,
"loss": 0.1942,
"step": 735
},
{
"epoch": 0.28145315487571704,
"grad_norm": 7.157943591816231,
"learning_rate": 9.19871234956512e-07,
"loss": 0.291,
"step": 736
},
{
"epoch": 0.28183556405353727,
"grad_norm": 6.540916189244941,
"learning_rate": 9.20060438593925e-07,
"loss": 0.1721,
"step": 737
},
{
"epoch": 0.28221797323135756,
"grad_norm": 6.142323188767702,
"learning_rate": 9.202493856839546e-07,
"loss": 0.3475,
"step": 738
},
{
"epoch": 0.28260038240917784,
"grad_norm": 5.9353717402798365,
"learning_rate": 9.204380769213804e-07,
"loss": 0.6243,
"step": 739
},
{
"epoch": 0.2829827915869981,
"grad_norm": 3.8146311322392403,
"learning_rate": 9.206265129981641e-07,
"loss": 0.659,
"step": 740
},
{
"epoch": 0.28336520076481836,
"grad_norm": 2.7360985255933583,
"learning_rate": 9.208146946034633e-07,
"loss": 0.3429,
"step": 741
},
{
"epoch": 0.28374760994263865,
"grad_norm": 3.501030089156713,
"learning_rate": 9.210026224236482e-07,
"loss": 0.4522,
"step": 742
},
{
"epoch": 0.2841300191204589,
"grad_norm": 2.9306435614494535,
"learning_rate": 9.211902971423152e-07,
"loss": 0.6043,
"step": 743
},
{
"epoch": 0.28451242829827916,
"grad_norm": 3.191075133909245,
"learning_rate": 9.213777194403025e-07,
"loss": 0.5403,
"step": 744
},
{
"epoch": 0.28489483747609945,
"grad_norm": 4.1194321016924285,
"learning_rate": 9.21564889995705e-07,
"loss": 0.1928,
"step": 745
},
{
"epoch": 0.2852772466539197,
"grad_norm": 2.9657019779034983,
"learning_rate": 9.217518094838887e-07,
"loss": 0.2501,
"step": 746
},
{
"epoch": 0.28565965583173997,
"grad_norm": 3.2172360609705466,
"learning_rate": 9.219384785775058e-07,
"loss": 0.4956,
"step": 747
},
{
"epoch": 0.28604206500956025,
"grad_norm": 2.7979749509223217,
"learning_rate": 9.221248979465084e-07,
"loss": 0.1676,
"step": 748
},
{
"epoch": 0.2864244741873805,
"grad_norm": 4.155212043281761,
"learning_rate": 9.223110682581634e-07,
"loss": 0.1349,
"step": 749
},
{
"epoch": 0.28680688336520077,
"grad_norm": 4.2952560426645,
"learning_rate": 9.224969901770675e-07,
"loss": 0.2733,
"step": 750
},
{
"epoch": 0.28718929254302106,
"grad_norm": 6.104755186628853,
"learning_rate": 9.226826643651602e-07,
"loss": 0.5484,
"step": 751
},
{
"epoch": 0.2875717017208413,
"grad_norm": 4.040718255733354,
"learning_rate": 9.228680914817387e-07,
"loss": 0.4754,
"step": 752
},
{
"epoch": 0.2879541108986616,
"grad_norm": 3.158374637813444,
"learning_rate": 9.230532721834717e-07,
"loss": 0.5389,
"step": 753
},
{
"epoch": 0.28833652007648186,
"grad_norm": 4.190667619708935,
"learning_rate": 9.232382071244136e-07,
"loss": 0.4755,
"step": 754
},
{
"epoch": 0.2887189292543021,
"grad_norm": 4.238531683629353,
"learning_rate": 9.234228969560181e-07,
"loss": 0.5469,
"step": 755
},
{
"epoch": 0.2891013384321224,
"grad_norm": 1.9615411931014568,
"learning_rate": 9.236073423271517e-07,
"loss": 0.3119,
"step": 756
},
{
"epoch": 0.28948374760994267,
"grad_norm": 2.6342224511339403,
"learning_rate": 9.237915438841085e-07,
"loss": 0.4063,
"step": 757
},
{
"epoch": 0.2898661567877629,
"grad_norm": 4.365004756614654,
"learning_rate": 9.239755022706225e-07,
"loss": 0.4622,
"step": 758
},
{
"epoch": 0.2902485659655832,
"grad_norm": 2.5054584994223874,
"learning_rate": 9.241592181278815e-07,
"loss": 0.312,
"step": 759
},
{
"epoch": 0.29063097514340347,
"grad_norm": 4.476758702956534,
"learning_rate": 9.243426920945409e-07,
"loss": 0.3864,
"step": 760
},
{
"epoch": 0.2910133843212237,
"grad_norm": 3.021955417639978,
"learning_rate": 9.245259248067367e-07,
"loss": 0.1333,
"step": 761
},
{
"epoch": 0.291395793499044,
"grad_norm": 4.690756315239575,
"learning_rate": 9.247089168980988e-07,
"loss": 0.1441,
"step": 762
},
{
"epoch": 0.2917782026768642,
"grad_norm": 7.829398959538781,
"learning_rate": 9.24891668999764e-07,
"loss": 0.2082,
"step": 763
},
{
"epoch": 0.2921606118546845,
"grad_norm": 5.258302209503068,
"learning_rate": 9.250741817403894e-07,
"loss": 0.6746,
"step": 764
},
{
"epoch": 0.2925430210325048,
"grad_norm": 4.864542185618209,
"learning_rate": 9.252564557461648e-07,
"loss": 0.6271,
"step": 765
},
{
"epoch": 0.292925430210325,
"grad_norm": 2.4464280384051205,
"learning_rate": 9.254384916408264e-07,
"loss": 0.3544,
"step": 766
},
{
"epoch": 0.2933078393881453,
"grad_norm": 1.737148919072545,
"learning_rate": 9.25620290045669e-07,
"loss": 0.4638,
"step": 767
},
{
"epoch": 0.2936902485659656,
"grad_norm": 3.298820030333467,
"learning_rate": 9.258018515795588e-07,
"loss": 0.3998,
"step": 768
},
{
"epoch": 0.2940726577437858,
"grad_norm": 2.7308884545045125,
"learning_rate": 9.25983176858946e-07,
"loss": 0.3025,
"step": 769
},
{
"epoch": 0.2944550669216061,
"grad_norm": 3.6917404104291522,
"learning_rate": 9.261642664978777e-07,
"loss": 0.3902,
"step": 770
},
{
"epoch": 0.2948374760994264,
"grad_norm": 3.438420536589146,
"learning_rate": 9.263451211080103e-07,
"loss": 0.1922,
"step": 771
},
{
"epoch": 0.29521988527724663,
"grad_norm": 2.690586446913425,
"learning_rate": 9.265257412986215e-07,
"loss": 0.1766,
"step": 772
},
{
"epoch": 0.2956022944550669,
"grad_norm": 3.889979297416141,
"learning_rate": 9.267061276766227e-07,
"loss": 0.1684,
"step": 773
},
{
"epoch": 0.2959847036328872,
"grad_norm": 5.210210883911897,
"learning_rate": 9.268862808465719e-07,
"loss": 0.143,
"step": 774
},
{
"epoch": 0.29636711281070743,
"grad_norm": 5.16822561719914,
"learning_rate": 9.270662014106851e-07,
"loss": 0.1774,
"step": 775
},
{
"epoch": 0.2967495219885277,
"grad_norm": 5.552930851914382,
"learning_rate": 9.272458899688485e-07,
"loss": 0.4315,
"step": 776
},
{
"epoch": 0.297131931166348,
"grad_norm": 3.574750216634461,
"learning_rate": 9.274253471186307e-07,
"loss": 0.6128,
"step": 777
},
{
"epoch": 0.29751434034416824,
"grad_norm": 3.457119249851013,
"learning_rate": 9.276045734552951e-07,
"loss": 0.6457,
"step": 778
},
{
"epoch": 0.2978967495219885,
"grad_norm": 3.1391020920925903,
"learning_rate": 9.277835695718105e-07,
"loss": 0.379,
"step": 779
},
{
"epoch": 0.2982791586998088,
"grad_norm": 3.245935686396853,
"learning_rate": 9.279623360588637e-07,
"loss": 0.3597,
"step": 780
},
{
"epoch": 0.29866156787762904,
"grad_norm": 4.948312629784024,
"learning_rate": 9.281408735048715e-07,
"loss": 0.5803,
"step": 781
},
{
"epoch": 0.29904397705544933,
"grad_norm": 3.5383041276653304,
"learning_rate": 9.283191824959918e-07,
"loss": 0.4158,
"step": 782
},
{
"epoch": 0.2994263862332696,
"grad_norm": 2.6961675642106813,
"learning_rate": 9.284972636161345e-07,
"loss": 0.234,
"step": 783
},
{
"epoch": 0.29980879541108985,
"grad_norm": 3.6581698862695102,
"learning_rate": 9.286751174469749e-07,
"loss": 0.2256,
"step": 784
},
{
"epoch": 0.30019120458891013,
"grad_norm": 3.658953979144392,
"learning_rate": 9.288527445679631e-07,
"loss": 0.2376,
"step": 785
},
{
"epoch": 0.3005736137667304,
"grad_norm": 4.144486839638947,
"learning_rate": 9.290301455563364e-07,
"loss": 0.1269,
"step": 786
},
{
"epoch": 0.30095602294455065,
"grad_norm": 4.947979200536853,
"learning_rate": 9.292073209871303e-07,
"loss": 0.1268,
"step": 787
},
{
"epoch": 0.30133843212237094,
"grad_norm": 6.4016710245636865,
"learning_rate": 9.293842714331897e-07,
"loss": 0.4139,
"step": 788
},
{
"epoch": 0.3017208413001912,
"grad_norm": 4.04470071597837,
"learning_rate": 9.295609974651798e-07,
"loss": 0.59,
"step": 789
},
{
"epoch": 0.30210325047801145,
"grad_norm": 4.723204474422228,
"learning_rate": 9.297374996515978e-07,
"loss": 0.6182,
"step": 790
},
{
"epoch": 0.30248565965583174,
"grad_norm": 3.057452343298742,
"learning_rate": 9.299137785587831e-07,
"loss": 0.537,
"step": 791
},
{
"epoch": 0.302868068833652,
"grad_norm": 2.19324306271355,
"learning_rate": 9.300898347509282e-07,
"loss": 0.6086,
"step": 792
},
{
"epoch": 0.30325047801147226,
"grad_norm": 2.9430732613303423,
"learning_rate": 9.302656687900906e-07,
"loss": 0.2686,
"step": 793
},
{
"epoch": 0.30363288718929254,
"grad_norm": 3.9081728665745197,
"learning_rate": 9.304412812362025e-07,
"loss": 0.444,
"step": 794
},
{
"epoch": 0.30401529636711283,
"grad_norm": 2.4089558237980917,
"learning_rate": 9.306166726470812e-07,
"loss": 0.3601,
"step": 795
},
{
"epoch": 0.30439770554493306,
"grad_norm": 3.4118547437366673,
"learning_rate": 9.307918435784414e-07,
"loss": 0.3465,
"step": 796
},
{
"epoch": 0.30478011472275335,
"grad_norm": 3.6235244641842153,
"learning_rate": 9.309667945839041e-07,
"loss": 0.1584,
"step": 797
},
{
"epoch": 0.30516252390057363,
"grad_norm": 3.437496948240833,
"learning_rate": 9.311415262150076e-07,
"loss": 0.2358,
"step": 798
},
{
"epoch": 0.30554493307839387,
"grad_norm": 3.4448784448801675,
"learning_rate": 9.313160390212184e-07,
"loss": 0.1215,
"step": 799
},
{
"epoch": 0.30592734225621415,
"grad_norm": 3.5191571611593258,
"learning_rate": 9.314903335499412e-07,
"loss": 0.1596,
"step": 800
},
{
"epoch": 0.30630975143403444,
"grad_norm": 6.999381174254604,
"learning_rate": 9.316644103465291e-07,
"loss": 0.5317,
"step": 801
},
{
"epoch": 0.30669216061185467,
"grad_norm": 3.616443532807995,
"learning_rate": 9.318382699542936e-07,
"loss": 0.3562,
"step": 802
},
{
"epoch": 0.30707456978967496,
"grad_norm": 2.3191912989439687,
"learning_rate": 9.320119129145158e-07,
"loss": 0.4623,
"step": 803
},
{
"epoch": 0.30745697896749524,
"grad_norm": 3.41795103232015,
"learning_rate": 9.321853397664554e-07,
"loss": 0.5882,
"step": 804
},
{
"epoch": 0.3078393881453155,
"grad_norm": 3.6543276697326195,
"learning_rate": 9.323585510473612e-07,
"loss": 0.4667,
"step": 805
},
{
"epoch": 0.30822179732313576,
"grad_norm": 2.31745390334883,
"learning_rate": 9.325315472924812e-07,
"loss": 0.2955,
"step": 806
},
{
"epoch": 0.30860420650095605,
"grad_norm": 4.031207417107943,
"learning_rate": 9.327043290350726e-07,
"loss": 0.3176,
"step": 807
},
{
"epoch": 0.3089866156787763,
"grad_norm": 4.097064125473531,
"learning_rate": 9.328768968064107e-07,
"loss": 0.3082,
"step": 808
},
{
"epoch": 0.30936902485659656,
"grad_norm": 2.4083338050005847,
"learning_rate": 9.330492511358006e-07,
"loss": 0.1999,
"step": 809
},
{
"epoch": 0.30975143403441685,
"grad_norm": 3.1727456086591035,
"learning_rate": 9.332213925505848e-07,
"loss": 0.1533,
"step": 810
},
{
"epoch": 0.3101338432122371,
"grad_norm": 4.600015117786278,
"learning_rate": 9.333933215761543e-07,
"loss": 0.2456,
"step": 811
},
{
"epoch": 0.31051625239005737,
"grad_norm": 7.192106412819565,
"learning_rate": 9.335650387359578e-07,
"loss": 0.2425,
"step": 812
},
{
"epoch": 0.31089866156787765,
"grad_norm": 7.02838320463472,
"learning_rate": 9.337365445515112e-07,
"loss": 0.2411,
"step": 813
},
{
"epoch": 0.3112810707456979,
"grad_norm": 4.490032601354969,
"learning_rate": 9.339078395424072e-07,
"loss": 0.6411,
"step": 814
},
{
"epoch": 0.31166347992351817,
"grad_norm": 3.2150438322932544,
"learning_rate": 9.340789242263248e-07,
"loss": 0.4324,
"step": 815
},
{
"epoch": 0.31204588910133846,
"grad_norm": 3.8721174778416825,
"learning_rate": 9.342497991190384e-07,
"loss": 0.3775,
"step": 816
},
{
"epoch": 0.3124282982791587,
"grad_norm": 4.4248550132662805,
"learning_rate": 9.344204647344277e-07,
"loss": 0.367,
"step": 817
},
{
"epoch": 0.312810707456979,
"grad_norm": 3.1897752906316335,
"learning_rate": 9.345909215844859e-07,
"loss": 0.2308,
"step": 818
},
{
"epoch": 0.31319311663479926,
"grad_norm": 3.450483597848423,
"learning_rate": 9.347611701793305e-07,
"loss": 0.4343,
"step": 819
},
{
"epoch": 0.3135755258126195,
"grad_norm": 2.481171179607701,
"learning_rate": 9.349312110272107e-07,
"loss": 0.3205,
"step": 820
},
{
"epoch": 0.3139579349904398,
"grad_norm": 2.788976245903436,
"learning_rate": 9.351010446345182e-07,
"loss": 0.2113,
"step": 821
},
{
"epoch": 0.31434034416826,
"grad_norm": 6.271641000165821,
"learning_rate": 9.352706715057951e-07,
"loss": 0.3858,
"step": 822
},
{
"epoch": 0.3147227533460803,
"grad_norm": 3.7436334607767705,
"learning_rate": 9.354400921437429e-07,
"loss": 0.163,
"step": 823
},
{
"epoch": 0.3151051625239006,
"grad_norm": 3.741192711347864,
"learning_rate": 9.356093070492322e-07,
"loss": 0.1277,
"step": 824
},
{
"epoch": 0.3154875717017208,
"grad_norm": 5.455032146299698,
"learning_rate": 9.357783167213107e-07,
"loss": 0.413,
"step": 825
},
{
"epoch": 0.3158699808795411,
"grad_norm": 7.935603838938422,
"learning_rate": 9.359471216572131e-07,
"loss": 0.5146,
"step": 826
},
{
"epoch": 0.3162523900573614,
"grad_norm": 5.691858654591456,
"learning_rate": 9.36115722352369e-07,
"loss": 0.7377,
"step": 827
},
{
"epoch": 0.3166347992351816,
"grad_norm": 3.392299537442618,
"learning_rate": 9.362841193004116e-07,
"loss": 0.5823,
"step": 828
},
{
"epoch": 0.3170172084130019,
"grad_norm": 3.2098505635401144,
"learning_rate": 9.36452312993187e-07,
"loss": 0.4236,
"step": 829
},
{
"epoch": 0.3173996175908222,
"grad_norm": 4.264147380292967,
"learning_rate": 9.36620303920762e-07,
"loss": 0.5482,
"step": 830
},
{
"epoch": 0.3177820267686424,
"grad_norm": 3.0817239702452555,
"learning_rate": 9.367880925714339e-07,
"loss": 0.3956,
"step": 831
},
{
"epoch": 0.3181644359464627,
"grad_norm": 3.747304837958074,
"learning_rate": 9.369556794317375e-07,
"loss": 0.581,
"step": 832
},
{
"epoch": 0.318546845124283,
"grad_norm": 3.7135952396817578,
"learning_rate": 9.371230649864548e-07,
"loss": 0.4357,
"step": 833
},
{
"epoch": 0.3189292543021032,
"grad_norm": 4.297254067086311,
"learning_rate": 9.372902497186226e-07,
"loss": 0.2272,
"step": 834
},
{
"epoch": 0.3193116634799235,
"grad_norm": 2.567065289328753,
"learning_rate": 9.374572341095416e-07,
"loss": 0.1196,
"step": 835
},
{
"epoch": 0.3196940726577438,
"grad_norm": 4.40650365653969,
"learning_rate": 9.376240186387841e-07,
"loss": 0.2121,
"step": 836
},
{
"epoch": 0.32007648183556403,
"grad_norm": 3.7109315892222665,
"learning_rate": 9.377906037842022e-07,
"loss": 0.248,
"step": 837
},
{
"epoch": 0.3204588910133843,
"grad_norm": 7.150674972152842,
"learning_rate": 9.379569900219368e-07,
"loss": 0.1896,
"step": 838
},
{
"epoch": 0.3208413001912046,
"grad_norm": 3.2211659077572907,
"learning_rate": 9.381231778264247e-07,
"loss": 0.5999,
"step": 839
},
{
"epoch": 0.32122370936902483,
"grad_norm": 4.1372208331105345,
"learning_rate": 9.382891676704079e-07,
"loss": 0.4782,
"step": 840
},
{
"epoch": 0.3216061185468451,
"grad_norm": 3.8547351452280036,
"learning_rate": 9.384549600249407e-07,
"loss": 0.3241,
"step": 841
},
{
"epoch": 0.3219885277246654,
"grad_norm": 2.3007511985383564,
"learning_rate": 9.386205553593977e-07,
"loss": 0.5596,
"step": 842
},
{
"epoch": 0.32237093690248564,
"grad_norm": 2.9696172150310205,
"learning_rate": 9.387859541414829e-07,
"loss": 0.6675,
"step": 843
},
{
"epoch": 0.3227533460803059,
"grad_norm": 3.513270153174881,
"learning_rate": 9.389511568372363e-07,
"loss": 0.4169,
"step": 844
},
{
"epoch": 0.3231357552581262,
"grad_norm": 2.6253018886813506,
"learning_rate": 9.391161639110424e-07,
"loss": 0.261,
"step": 845
},
{
"epoch": 0.32351816443594644,
"grad_norm": 3.17710482897548,
"learning_rate": 9.392809758256383e-07,
"loss": 0.2942,
"step": 846
},
{
"epoch": 0.32390057361376673,
"grad_norm": 7.235980455126496,
"learning_rate": 9.39445593042121e-07,
"loss": 0.1672,
"step": 847
},
{
"epoch": 0.324282982791587,
"grad_norm": 3.4921857249278907,
"learning_rate": 9.396100160199549e-07,
"loss": 0.2866,
"step": 848
},
{
"epoch": 0.32466539196940725,
"grad_norm": 3.557020021551161,
"learning_rate": 9.397742452169806e-07,
"loss": 0.1083,
"step": 849
},
{
"epoch": 0.32504780114722753,
"grad_norm": 4.516592588673517,
"learning_rate": 9.399382810894211e-07,
"loss": 0.1744,
"step": 850
},
{
"epoch": 0.3254302103250478,
"grad_norm": 4.88414591168416,
"learning_rate": 9.401021240918907e-07,
"loss": 0.6063,
"step": 851
},
{
"epoch": 0.32581261950286805,
"grad_norm": 3.997072937025769,
"learning_rate": 9.402657746774017e-07,
"loss": 0.5326,
"step": 852
},
{
"epoch": 0.32619502868068834,
"grad_norm": 2.682499513937132,
"learning_rate": 9.404292332973727e-07,
"loss": 0.7379,
"step": 853
},
{
"epoch": 0.3265774378585086,
"grad_norm": 1.6899059945545842,
"learning_rate": 9.405925004016348e-07,
"loss": 0.4393,
"step": 854
},
{
"epoch": 0.32695984703632885,
"grad_norm": 3.3084039374696004,
"learning_rate": 9.407555764384405e-07,
"loss": 0.5194,
"step": 855
},
{
"epoch": 0.32734225621414914,
"grad_norm": 4.0873745493638225,
"learning_rate": 9.409184618544701e-07,
"loss": 0.5545,
"step": 856
},
{
"epoch": 0.3277246653919694,
"grad_norm": 2.995847371103419,
"learning_rate": 9.410811570948394e-07,
"loss": 0.2915,
"step": 857
},
{
"epoch": 0.32810707456978966,
"grad_norm": 2.946807077379506,
"learning_rate": 9.412436626031069e-07,
"loss": 0.2918,
"step": 858
},
{
"epoch": 0.32848948374760994,
"grad_norm": 3.597120744279483,
"learning_rate": 9.414059788212811e-07,
"loss": 0.4846,
"step": 859
},
{
"epoch": 0.32887189292543023,
"grad_norm": 3.694052096178697,
"learning_rate": 9.415681061898281e-07,
"loss": 0.3207,
"step": 860
},
{
"epoch": 0.32925430210325046,
"grad_norm": 3.617068591542446,
"learning_rate": 9.417300451476775e-07,
"loss": 0.1891,
"step": 861
},
{
"epoch": 0.32963671128107075,
"grad_norm": 3.738340690243869,
"learning_rate": 9.418917961322313e-07,
"loss": 0.1135,
"step": 862
},
{
"epoch": 0.33001912045889104,
"grad_norm": 5.789680231781311,
"learning_rate": 9.420533595793696e-07,
"loss": 0.2475,
"step": 863
},
{
"epoch": 0.33040152963671127,
"grad_norm": 5.945408743882622,
"learning_rate": 9.422147359234583e-07,
"loss": 0.6057,
"step": 864
},
{
"epoch": 0.33078393881453155,
"grad_norm": 4.560439389124144,
"learning_rate": 9.423759255973562e-07,
"loss": 0.5982,
"step": 865
},
{
"epoch": 0.33116634799235184,
"grad_norm": 4.703064370002157,
"learning_rate": 9.425369290324213e-07,
"loss": 0.4356,
"step": 866
},
{
"epoch": 0.33154875717017207,
"grad_norm": 2.6385904634172532,
"learning_rate": 9.426977466585182e-07,
"loss": 0.3505,
"step": 867
},
{
"epoch": 0.33193116634799236,
"grad_norm": 3.0721824329168745,
"learning_rate": 9.428583789040255e-07,
"loss": 0.4258,
"step": 868
},
{
"epoch": 0.33231357552581264,
"grad_norm": 2.8782761193446076,
"learning_rate": 9.43018826195841e-07,
"loss": 0.3341,
"step": 869
},
{
"epoch": 0.3326959847036329,
"grad_norm": 4.241642307493842,
"learning_rate": 9.431790889593908e-07,
"loss": 0.2586,
"step": 870
},
{
"epoch": 0.33307839388145316,
"grad_norm": 2.5946874188452203,
"learning_rate": 9.433391676186339e-07,
"loss": 0.452,
"step": 871
},
{
"epoch": 0.33346080305927345,
"grad_norm": 2.8512966084297795,
"learning_rate": 9.434990625960707e-07,
"loss": 0.2671,
"step": 872
},
{
"epoch": 0.3338432122370937,
"grad_norm": 5.01618170099747,
"learning_rate": 9.436587743127481e-07,
"loss": 0.3723,
"step": 873
},
{
"epoch": 0.33422562141491396,
"grad_norm": 4.551976020065996,
"learning_rate": 9.438183031882673e-07,
"loss": 0.2959,
"step": 874
},
{
"epoch": 0.33460803059273425,
"grad_norm": 5.27782676852505,
"learning_rate": 9.439776496407904e-07,
"loss": 0.338,
"step": 875
},
{
"epoch": 0.3349904397705545,
"grad_norm": 5.48133248383455,
"learning_rate": 9.441368140870459e-07,
"loss": 0.29,
"step": 876
},
{
"epoch": 0.33537284894837477,
"grad_norm": 3.56000270650525,
"learning_rate": 9.442957969423364e-07,
"loss": 0.5127,
"step": 877
},
{
"epoch": 0.33575525812619506,
"grad_norm": 2.825018003288225,
"learning_rate": 9.444545986205447e-07,
"loss": 0.3841,
"step": 878
},
{
"epoch": 0.3361376673040153,
"grad_norm": 2.008479501093062,
"learning_rate": 9.446132195341398e-07,
"loss": 0.4476,
"step": 879
},
{
"epoch": 0.3365200764818356,
"grad_norm": 2.281587131619646,
"learning_rate": 9.447716600941844e-07,
"loss": 0.4282,
"step": 880
},
{
"epoch": 0.33690248565965586,
"grad_norm": 4.628300494383288,
"learning_rate": 9.449299207103402e-07,
"loss": 0.3542,
"step": 881
},
{
"epoch": 0.3372848948374761,
"grad_norm": 2.3056502061077704,
"learning_rate": 9.450880017908747e-07,
"loss": 0.4025,
"step": 882
},
{
"epoch": 0.3376673040152964,
"grad_norm": 2.4249371274682714,
"learning_rate": 9.452459037426675e-07,
"loss": 0.4326,
"step": 883
},
{
"epoch": 0.3380497131931166,
"grad_norm": 2.783489225974867,
"learning_rate": 9.454036269712172e-07,
"loss": 0.325,
"step": 884
},
{
"epoch": 0.3384321223709369,
"grad_norm": 4.7547218544653145,
"learning_rate": 9.455611718806462e-07,
"loss": 0.2726,
"step": 885
},
{
"epoch": 0.3388145315487572,
"grad_norm": 2.723020334953059,
"learning_rate": 9.457185388737084e-07,
"loss": 0.2322,
"step": 886
},
{
"epoch": 0.3391969407265774,
"grad_norm": 3.7076447665554233,
"learning_rate": 9.458757283517939e-07,
"loss": 0.1403,
"step": 887
},
{
"epoch": 0.3395793499043977,
"grad_norm": 6.507640163159305,
"learning_rate": 9.460327407149374e-07,
"loss": 0.2687,
"step": 888
},
{
"epoch": 0.339961759082218,
"grad_norm": 3.3387596467856,
"learning_rate": 9.461895763618216e-07,
"loss": 0.4304,
"step": 889
},
{
"epoch": 0.3403441682600382,
"grad_norm": 4.245030358591246,
"learning_rate": 9.463462356897853e-07,
"loss": 0.7139,
"step": 890
},
{
"epoch": 0.3407265774378585,
"grad_norm": 3.128485300081678,
"learning_rate": 9.46502719094828e-07,
"loss": 0.489,
"step": 891
},
{
"epoch": 0.3411089866156788,
"grad_norm": 2.747736953361468,
"learning_rate": 9.466590269716174e-07,
"loss": 0.4317,
"step": 892
},
{
"epoch": 0.341491395793499,
"grad_norm": 2.683976458450521,
"learning_rate": 9.468151597134941e-07,
"loss": 0.3164,
"step": 893
},
{
"epoch": 0.3418738049713193,
"grad_norm": 2.566947427066389,
"learning_rate": 9.469711177124784e-07,
"loss": 0.2012,
"step": 894
},
{
"epoch": 0.3422562141491396,
"grad_norm": 4.025599104560061,
"learning_rate": 9.471269013592754e-07,
"loss": 0.3256,
"step": 895
},
{
"epoch": 0.3426386233269598,
"grad_norm": 3.243968061052379,
"learning_rate": 9.472825110432816e-07,
"loss": 0.3572,
"step": 896
},
{
"epoch": 0.3430210325047801,
"grad_norm": 2.8867854832019284,
"learning_rate": 9.474379471525903e-07,
"loss": 0.2392,
"step": 897
},
{
"epoch": 0.3434034416826004,
"grad_norm": 4.063409322059927,
"learning_rate": 9.475932100739977e-07,
"loss": 0.1856,
"step": 898
},
{
"epoch": 0.3437858508604206,
"grad_norm": 4.12406679636957,
"learning_rate": 9.477483001930082e-07,
"loss": 0.1849,
"step": 899
},
{
"epoch": 0.3441682600382409,
"grad_norm": 5.161388818893121,
"learning_rate": 9.47903217893841e-07,
"loss": 0.1993,
"step": 900
},
{
"epoch": 0.3445506692160612,
"grad_norm": 5.069430658098168,
"learning_rate": 9.480579635594347e-07,
"loss": 0.6176,
"step": 901
},
{
"epoch": 0.34493307839388143,
"grad_norm": 4.224126032437366,
"learning_rate": 9.48212537571454e-07,
"loss": 0.651,
"step": 902
},
{
"epoch": 0.3453154875717017,
"grad_norm": 4.250506651354976,
"learning_rate": 9.483669403102948e-07,
"loss": 0.5044,
"step": 903
},
{
"epoch": 0.345697896749522,
"grad_norm": 2.465775931686004,
"learning_rate": 9.485211721550897e-07,
"loss": 0.3392,
"step": 904
},
{
"epoch": 0.34608030592734224,
"grad_norm": 2.3975292561230748,
"learning_rate": 9.48675233483714e-07,
"loss": 0.3344,
"step": 905
},
{
"epoch": 0.3464627151051625,
"grad_norm": 3.0526375294539094,
"learning_rate": 9.488291246727914e-07,
"loss": 0.4743,
"step": 906
},
{
"epoch": 0.3468451242829828,
"grad_norm": 2.336490425323715,
"learning_rate": 9.489828460976984e-07,
"loss": 0.5388,
"step": 907
},
{
"epoch": 0.34722753346080304,
"grad_norm": 2.5558834717862426,
"learning_rate": 9.491363981325713e-07,
"loss": 0.2754,
"step": 908
},
{
"epoch": 0.3476099426386233,
"grad_norm": 4.177412950760512,
"learning_rate": 9.492897811503105e-07,
"loss": 0.3361,
"step": 909
},
{
"epoch": 0.3479923518164436,
"grad_norm": 4.640426888395395,
"learning_rate": 9.494429955225867e-07,
"loss": 0.4094,
"step": 910
},
{
"epoch": 0.34837476099426384,
"grad_norm": 3.038139771953519,
"learning_rate": 9.495960416198456e-07,
"loss": 0.1049,
"step": 911
},
{
"epoch": 0.34875717017208413,
"grad_norm": 3.971266181679893,
"learning_rate": 9.497489198113142e-07,
"loss": 0.1444,
"step": 912
},
{
"epoch": 0.3491395793499044,
"grad_norm": 5.104859949894045,
"learning_rate": 9.499016304650052e-07,
"loss": 0.3623,
"step": 913
},
{
"epoch": 0.34952198852772465,
"grad_norm": 6.547232752382276,
"learning_rate": 9.500541739477227e-07,
"loss": 0.3693,
"step": 914
},
{
"epoch": 0.34990439770554493,
"grad_norm": 3.6828872796074594,
"learning_rate": 9.502065506250678e-07,
"loss": 0.5371,
"step": 915
},
{
"epoch": 0.3502868068833652,
"grad_norm": 3.6765337020842113,
"learning_rate": 9.503587608614433e-07,
"loss": 0.6631,
"step": 916
},
{
"epoch": 0.35066921606118545,
"grad_norm": 2.8021830903817624,
"learning_rate": 9.505108050200593e-07,
"loss": 0.3992,
"step": 917
},
{
"epoch": 0.35105162523900574,
"grad_norm": 3.703896615161285,
"learning_rate": 9.506626834629382e-07,
"loss": 0.5016,
"step": 918
},
{
"epoch": 0.351434034416826,
"grad_norm": 3.866072430385559,
"learning_rate": 9.508143965509201e-07,
"loss": 0.4969,
"step": 919
},
{
"epoch": 0.35181644359464626,
"grad_norm": 2.872815172718489,
"learning_rate": 9.509659446436679e-07,
"loss": 0.5259,
"step": 920
},
{
"epoch": 0.35219885277246654,
"grad_norm": 2.5012410897968893,
"learning_rate": 9.511173280996717e-07,
"loss": 0.3576,
"step": 921
},
{
"epoch": 0.35258126195028683,
"grad_norm": 4.912924727310679,
"learning_rate": 9.512685472762555e-07,
"loss": 0.2694,
"step": 922
},
{
"epoch": 0.35296367112810706,
"grad_norm": 3.868625812663427,
"learning_rate": 9.514196025295804e-07,
"loss": 0.1522,
"step": 923
},
{
"epoch": 0.35334608030592735,
"grad_norm": 3.2232447971096962,
"learning_rate": 9.515704942146511e-07,
"loss": 0.0933,
"step": 924
},
{
"epoch": 0.35372848948374763,
"grad_norm": 4.741836962242508,
"learning_rate": 9.5172122268532e-07,
"loss": 0.199,
"step": 925
},
{
"epoch": 0.35411089866156786,
"grad_norm": 11.242952979498702,
"learning_rate": 9.518717882942924e-07,
"loss": 0.3926,
"step": 926
},
{
"epoch": 0.35449330783938815,
"grad_norm": 3.748092420499065,
"learning_rate": 9.520221913931318e-07,
"loss": 0.3953,
"step": 927
},
{
"epoch": 0.35487571701720844,
"grad_norm": 2.366885628667578,
"learning_rate": 9.521724323322646e-07,
"loss": 0.574,
"step": 928
},
{
"epoch": 0.35525812619502867,
"grad_norm": 3.8608011552795505,
"learning_rate": 9.523225114609844e-07,
"loss": 0.3235,
"step": 929
},
{
"epoch": 0.35564053537284895,
"grad_norm": 3.9321756846858174,
"learning_rate": 9.524724291274583e-07,
"loss": 0.4958,
"step": 930
},
{
"epoch": 0.35602294455066924,
"grad_norm": 3.548843845632833,
"learning_rate": 9.526221856787305e-07,
"loss": 0.2713,
"step": 931
},
{
"epoch": 0.35640535372848947,
"grad_norm": 2.545063146620941,
"learning_rate": 9.527717814607271e-07,
"loss": 0.3894,
"step": 932
},
{
"epoch": 0.35678776290630976,
"grad_norm": 2.8256888672163294,
"learning_rate": 9.529212168182619e-07,
"loss": 0.3203,
"step": 933
},
{
"epoch": 0.35717017208413004,
"grad_norm": 3.8043380670836955,
"learning_rate": 9.530704920950403e-07,
"loss": 0.2826,
"step": 934
},
{
"epoch": 0.3575525812619503,
"grad_norm": 3.1976715557254938,
"learning_rate": 9.532196076336643e-07,
"loss": 0.2002,
"step": 935
},
{
"epoch": 0.35793499043977056,
"grad_norm": 4.822091588850192,
"learning_rate": 9.533685637756371e-07,
"loss": 0.3646,
"step": 936
},
{
"epoch": 0.35831739961759085,
"grad_norm": 6.43557710539182,
"learning_rate": 9.535173608613685e-07,
"loss": 0.1462,
"step": 937
},
{
"epoch": 0.3586998087954111,
"grad_norm": 5.112594663307877,
"learning_rate": 9.536659992301782e-07,
"loss": 0.3242,
"step": 938
},
{
"epoch": 0.35908221797323137,
"grad_norm": 4.220278314750709,
"learning_rate": 9.538144792203018e-07,
"loss": 0.4659,
"step": 939
},
{
"epoch": 0.35946462715105165,
"grad_norm": 3.2606053079423964,
"learning_rate": 9.539628011688945e-07,
"loss": 0.623,
"step": 940
},
{
"epoch": 0.3598470363288719,
"grad_norm": 2.972337021398215,
"learning_rate": 9.541109654120364e-07,
"loss": 0.4305,
"step": 941
},
{
"epoch": 0.36022944550669217,
"grad_norm": 3.5581539481074476,
"learning_rate": 9.542589722847364e-07,
"loss": 0.559,
"step": 942
},
{
"epoch": 0.3606118546845124,
"grad_norm": 4.023866500945532,
"learning_rate": 9.544068221209373e-07,
"loss": 0.3209,
"step": 943
},
{
"epoch": 0.3609942638623327,
"grad_norm": 2.35951869097576,
"learning_rate": 9.5455451525352e-07,
"loss": 0.4073,
"step": 944
},
{
"epoch": 0.361376673040153,
"grad_norm": 4.300513152619637,
"learning_rate": 9.547020520143076e-07,
"loss": 0.2415,
"step": 945
},
{
"epoch": 0.3617590822179732,
"grad_norm": 2.523667835789399,
"learning_rate": 9.548494327340713e-07,
"loss": 0.2748,
"step": 946
},
{
"epoch": 0.3621414913957935,
"grad_norm": 3.440650190075816,
"learning_rate": 9.54996657742533e-07,
"loss": 0.3607,
"step": 947
},
{
"epoch": 0.3625239005736138,
"grad_norm": 3.994097885238933,
"learning_rate": 9.551437273683711e-07,
"loss": 0.276,
"step": 948
},
{
"epoch": 0.362906309751434,
"grad_norm": 3.940188141218107,
"learning_rate": 9.552906419392247e-07,
"loss": 0.1379,
"step": 949
},
{
"epoch": 0.3632887189292543,
"grad_norm": 4.636109489642276,
"learning_rate": 9.554374017816967e-07,
"loss": 0.2433,
"step": 950
},
{
"epoch": 0.3636711281070746,
"grad_norm": 7.117983659417351,
"learning_rate": 9.555840072213603e-07,
"loss": 0.515,
"step": 951
},
{
"epoch": 0.3640535372848948,
"grad_norm": 4.1641341842006385,
"learning_rate": 9.557304585827614e-07,
"loss": 0.6473,
"step": 952
},
{
"epoch": 0.3644359464627151,
"grad_norm": 3.9707883871656287,
"learning_rate": 9.55876756189424e-07,
"loss": 0.6406,
"step": 953
},
{
"epoch": 0.3648183556405354,
"grad_norm": 3.0221837324920293,
"learning_rate": 9.560229003638547e-07,
"loss": 0.4578,
"step": 954
},
{
"epoch": 0.3652007648183556,
"grad_norm": 2.78359928992419,
"learning_rate": 9.561688914275451e-07,
"loss": 0.4396,
"step": 955
},
{
"epoch": 0.3655831739961759,
"grad_norm": 3.451375941237979,
"learning_rate": 9.56314729700979e-07,
"loss": 0.5369,
"step": 956
},
{
"epoch": 0.3659655831739962,
"grad_norm": 4.276913193584142,
"learning_rate": 9.56460415503634e-07,
"loss": 0.3295,
"step": 957
},
{
"epoch": 0.3663479923518164,
"grad_norm": 3.4730854605922308,
"learning_rate": 9.56605949153987e-07,
"loss": 0.4103,
"step": 958
},
{
"epoch": 0.3667304015296367,
"grad_norm": 4.399869153938256,
"learning_rate": 9.56751330969518e-07,
"loss": 0.24,
"step": 959
},
{
"epoch": 0.367112810707457,
"grad_norm": 3.7183967230189814,
"learning_rate": 9.568965612667146e-07,
"loss": 0.2586,
"step": 960
},
{
"epoch": 0.3674952198852772,
"grad_norm": 3.0025703227552922,
"learning_rate": 9.57041640361076e-07,
"loss": 0.1275,
"step": 961
},
{
"epoch": 0.3678776290630975,
"grad_norm": 2.641218130097246,
"learning_rate": 9.571865685671162e-07,
"loss": 0.1138,
"step": 962
},
{
"epoch": 0.3682600382409178,
"grad_norm": 6.87125641596111,
"learning_rate": 9.5733134619837e-07,
"loss": 0.3016,
"step": 963
},
{
"epoch": 0.36864244741873803,
"grad_norm": 2.9957571226002058,
"learning_rate": 9.574759735673949e-07,
"loss": 0.6492,
"step": 964
},
{
"epoch": 0.3690248565965583,
"grad_norm": 2.803405723846682,
"learning_rate": 9.576204509857772e-07,
"loss": 0.529,
"step": 965
},
{
"epoch": 0.3694072657743786,
"grad_norm": 3.5725217699224485,
"learning_rate": 9.577647787641344e-07,
"loss": 0.5717,
"step": 966
},
{
"epoch": 0.36978967495219883,
"grad_norm": 3.079030324956933,
"learning_rate": 9.5790895721212e-07,
"loss": 0.2471,
"step": 967
},
{
"epoch": 0.3701720841300191,
"grad_norm": 3.102122988851968,
"learning_rate": 9.580529866384277e-07,
"loss": 0.2238,
"step": 968
},
{
"epoch": 0.3705544933078394,
"grad_norm": 2.5352702298867085,
"learning_rate": 9.58196867350794e-07,
"loss": 0.2231,
"step": 969
},
{
"epoch": 0.37093690248565964,
"grad_norm": 3.6590706142420597,
"learning_rate": 9.583405996560044e-07,
"loss": 0.4481,
"step": 970
},
{
"epoch": 0.3713193116634799,
"grad_norm": 2.4556048062747413,
"learning_rate": 9.584841838598948e-07,
"loss": 0.3815,
"step": 971
},
{
"epoch": 0.3717017208413002,
"grad_norm": 3.831446860702458,
"learning_rate": 9.586276202673582e-07,
"loss": 0.1587,
"step": 972
},
{
"epoch": 0.37208413001912044,
"grad_norm": 4.014302670550158,
"learning_rate": 9.587709091823455e-07,
"loss": 0.2652,
"step": 973
},
{
"epoch": 0.3724665391969407,
"grad_norm": 4.001061298719788,
"learning_rate": 9.589140509078721e-07,
"loss": 0.1145,
"step": 974
},
{
"epoch": 0.372848948374761,
"grad_norm": 5.591551790985613,
"learning_rate": 9.590570457460196e-07,
"loss": 0.1634,
"step": 975
},
{
"epoch": 0.37323135755258124,
"grad_norm": 7.88177658850859,
"learning_rate": 9.591998939979414e-07,
"loss": 0.5488,
"step": 976
},
{
"epoch": 0.37361376673040153,
"grad_norm": 3.544611944096069,
"learning_rate": 9.593425959638656e-07,
"loss": 0.564,
"step": 977
},
{
"epoch": 0.3739961759082218,
"grad_norm": 3.1957798548422,
"learning_rate": 9.594851519430982e-07,
"loss": 0.5497,
"step": 978
},
{
"epoch": 0.37437858508604205,
"grad_norm": 2.273287109434571,
"learning_rate": 9.596275622340283e-07,
"loss": 0.4016,
"step": 979
},
{
"epoch": 0.37476099426386233,
"grad_norm": 5.093270308485771,
"learning_rate": 9.597698271341309e-07,
"loss": 0.3785,
"step": 980
},
{
"epoch": 0.3751434034416826,
"grad_norm": 3.8876978882103375,
"learning_rate": 9.599119469399705e-07,
"loss": 0.38,
"step": 981
},
{
"epoch": 0.37552581261950285,
"grad_norm": 3.890440051726995,
"learning_rate": 9.600539219472053e-07,
"loss": 0.3075,
"step": 982
},
{
"epoch": 0.37590822179732314,
"grad_norm": 3.4391686290984627,
"learning_rate": 9.60195752450591e-07,
"loss": 0.5038,
"step": 983
},
{
"epoch": 0.3762906309751434,
"grad_norm": 5.500985837495509,
"learning_rate": 9.603374387439842e-07,
"loss": 0.3613,
"step": 984
},
{
"epoch": 0.37667304015296366,
"grad_norm": 5.084852064713341,
"learning_rate": 9.604789811203456e-07,
"loss": 0.153,
"step": 985
},
{
"epoch": 0.37705544933078394,
"grad_norm": 3.5089018559660863,
"learning_rate": 9.606203798717442e-07,
"loss": 0.2237,
"step": 986
},
{
"epoch": 0.37743785850860423,
"grad_norm": 5.30939060543432,
"learning_rate": 9.607616352893613e-07,
"loss": 0.304,
"step": 987
},
{
"epoch": 0.37782026768642446,
"grad_norm": 10.545777214279783,
"learning_rate": 9.60902747663493e-07,
"loss": 0.3001,
"step": 988
},
{
"epoch": 0.37820267686424475,
"grad_norm": 3.1998032628617197,
"learning_rate": 9.610437172835547e-07,
"loss": 0.4488,
"step": 989
},
{
"epoch": 0.37858508604206503,
"grad_norm": 3.828534038255384,
"learning_rate": 9.611845444380842e-07,
"loss": 0.5062,
"step": 990
},
{
"epoch": 0.37896749521988526,
"grad_norm": 3.1512350900517463,
"learning_rate": 9.613252294147454e-07,
"loss": 0.4848,
"step": 991
},
{
"epoch": 0.37934990439770555,
"grad_norm": 2.554872465291332,
"learning_rate": 9.61465772500332e-07,
"loss": 0.4056,
"step": 992
},
{
"epoch": 0.37973231357552584,
"grad_norm": 2.996642777372816,
"learning_rate": 9.616061739807709e-07,
"loss": 0.6107,
"step": 993
},
{
"epoch": 0.38011472275334607,
"grad_norm": 2.9023946806391008,
"learning_rate": 9.617464341411247e-07,
"loss": 0.3574,
"step": 994
},
{
"epoch": 0.38049713193116635,
"grad_norm": 2.5186035340842863,
"learning_rate": 9.618865532655972e-07,
"loss": 0.3863,
"step": 995
},
{
"epoch": 0.38087954110898664,
"grad_norm": 2.5047887713828496,
"learning_rate": 9.620265316375355e-07,
"loss": 0.342,
"step": 996
},
{
"epoch": 0.38126195028680687,
"grad_norm": 4.023505526958285,
"learning_rate": 9.621663695394329e-07,
"loss": 0.1651,
"step": 997
},
{
"epoch": 0.38164435946462716,
"grad_norm": 2.9972760231516817,
"learning_rate": 9.623060672529342e-07,
"loss": 0.1481,
"step": 998
},
{
"epoch": 0.38202676864244745,
"grad_norm": 3.7625935489012625,
"learning_rate": 9.624456250588371e-07,
"loss": 0.1253,
"step": 999
},
{
"epoch": 0.3824091778202677,
"grad_norm": 4.143639960345454,
"learning_rate": 9.625850432370972e-07,
"loss": 0.2301,
"step": 1000
},
{
"epoch": 0.3824091778202677,
"eval_runtime": 822.5233,
"eval_samples_per_second": 1.865,
"eval_steps_per_second": 0.467,
"step": 1000
},
{
"epoch": 0.38279158699808796,
"grad_norm": 13.168096794653332,
"learning_rate": 9.627243220668299e-07,
"loss": 0.3114,
"step": 1001
},
{
"epoch": 0.38317399617590825,
"grad_norm": 4.6670873089084814,
"learning_rate": 9.62863461826315e-07,
"loss": 0.5301,
"step": 1002
},
{
"epoch": 0.3835564053537285,
"grad_norm": 3.4543609478607573,
"learning_rate": 9.630024627929996e-07,
"loss": 0.445,
"step": 1003
},
{
"epoch": 0.38393881453154877,
"grad_norm": 2.9601279890896595,
"learning_rate": 9.631413252435012e-07,
"loss": 0.4941,
"step": 1004
},
{
"epoch": 0.384321223709369,
"grad_norm": 4.258867325618021,
"learning_rate": 9.632800494536112e-07,
"loss": 0.4981,
"step": 1005
},
{
"epoch": 0.3847036328871893,
"grad_norm": 2.933275959869867,
"learning_rate": 9.634186356982979e-07,
"loss": 0.3513,
"step": 1006
},
{
"epoch": 0.38508604206500957,
"grad_norm": 2.9809046835212243,
"learning_rate": 9.635570842517104e-07,
"loss": 0.4706,
"step": 1007
},
{
"epoch": 0.3854684512428298,
"grad_norm": 2.8025353975808414,
"learning_rate": 9.636953953871814e-07,
"loss": 0.2774,
"step": 1008
},
{
"epoch": 0.3858508604206501,
"grad_norm": 3.7234322698750906,
"learning_rate": 9.638335693772303e-07,
"loss": 0.3433,
"step": 1009
},
{
"epoch": 0.3862332695984704,
"grad_norm": 3.5164420979708093,
"learning_rate": 9.639716064935667e-07,
"loss": 0.3517,
"step": 1010
},
{
"epoch": 0.3866156787762906,
"grad_norm": 3.5137337577930254,
"learning_rate": 9.641095070070938e-07,
"loss": 0.0892,
"step": 1011
},
{
"epoch": 0.3869980879541109,
"grad_norm": 2.5491696183768577,
"learning_rate": 9.64247271187911e-07,
"loss": 0.1025,
"step": 1012
},
{
"epoch": 0.3873804971319312,
"grad_norm": 5.791968724266462,
"learning_rate": 9.643848993053177e-07,
"loss": 0.4488,
"step": 1013
},
{
"epoch": 0.3877629063097514,
"grad_norm": 4.08149153504387,
"learning_rate": 9.64522391627816e-07,
"loss": 0.7407,
"step": 1014
},
{
"epoch": 0.3881453154875717,
"grad_norm": 2.888440435602648,
"learning_rate": 9.646597484231137e-07,
"loss": 0.3919,
"step": 1015
},
{
"epoch": 0.388527724665392,
"grad_norm": 2.506209672332284,
"learning_rate": 9.647969699581283e-07,
"loss": 0.3635,
"step": 1016
},
{
"epoch": 0.3889101338432122,
"grad_norm": 4.320370554749521,
"learning_rate": 9.649340564989894e-07,
"loss": 0.4528,
"step": 1017
},
{
"epoch": 0.3892925430210325,
"grad_norm": 3.345833672839789,
"learning_rate": 9.650710083110417e-07,
"loss": 0.3671,
"step": 1018
},
{
"epoch": 0.3896749521988528,
"grad_norm": 3.1623782349315905,
"learning_rate": 9.652078256588485e-07,
"loss": 0.3302,
"step": 1019
},
{
"epoch": 0.390057361376673,
"grad_norm": 2.825148728768212,
"learning_rate": 9.653445088061943e-07,
"loss": 0.3511,
"step": 1020
},
{
"epoch": 0.3904397705544933,
"grad_norm": 3.1126792641610357,
"learning_rate": 9.654810580160888e-07,
"loss": 0.3865,
"step": 1021
},
{
"epoch": 0.3908221797323136,
"grad_norm": 3.0898730714277214,
"learning_rate": 9.656174735507689e-07,
"loss": 0.1243,
"step": 1022
},
{
"epoch": 0.3912045889101338,
"grad_norm": 3.4606426449243703,
"learning_rate": 9.657537556717016e-07,
"loss": 0.1593,
"step": 1023
},
{
"epoch": 0.3915869980879541,
"grad_norm": 3.169414599544937,
"learning_rate": 9.658899046395884e-07,
"loss": 0.1735,
"step": 1024
},
{
"epoch": 0.3919694072657744,
"grad_norm": 3.807031179859044,
"learning_rate": 9.660259207143668e-07,
"loss": 0.2093,
"step": 1025
},
{
"epoch": 0.3923518164435946,
"grad_norm": 7.122380628459166,
"learning_rate": 9.66161804155214e-07,
"loss": 0.4193,
"step": 1026
},
{
"epoch": 0.3927342256214149,
"grad_norm": 3.9833003969839265,
"learning_rate": 9.6629755522055e-07,
"loss": 0.7618,
"step": 1027
},
{
"epoch": 0.3931166347992352,
"grad_norm": 3.7067925041665486,
"learning_rate": 9.664331741680397e-07,
"loss": 0.3657,
"step": 1028
},
{
"epoch": 0.39349904397705543,
"grad_norm": 3.1871963057109762,
"learning_rate": 9.665686612545975e-07,
"loss": 0.4019,
"step": 1029
},
{
"epoch": 0.3938814531548757,
"grad_norm": 3.2304683071814235,
"learning_rate": 9.66704016736388e-07,
"loss": 0.5361,
"step": 1030
},
{
"epoch": 0.394263862332696,
"grad_norm": 5.435379491785883,
"learning_rate": 9.668392408688305e-07,
"loss": 0.3405,
"step": 1031
},
{
"epoch": 0.39464627151051623,
"grad_norm": 2.7119694485364176,
"learning_rate": 9.669743339066014e-07,
"loss": 0.368,
"step": 1032
},
{
"epoch": 0.3950286806883365,
"grad_norm": 2.539139121440049,
"learning_rate": 9.671092961036375e-07,
"loss": 0.2497,
"step": 1033
},
{
"epoch": 0.3954110898661568,
"grad_norm": 5.451613252184311,
"learning_rate": 9.672441277131377e-07,
"loss": 0.3053,
"step": 1034
},
{
"epoch": 0.39579349904397704,
"grad_norm": 3.4615202471262028,
"learning_rate": 9.673788289875675e-07,
"loss": 0.2826,
"step": 1035
},
{
"epoch": 0.3961759082217973,
"grad_norm": 3.0198202253984565,
"learning_rate": 9.675134001786604e-07,
"loss": 0.1403,
"step": 1036
},
{
"epoch": 0.3965583173996176,
"grad_norm": 3.1099402139197587,
"learning_rate": 9.676478415374213e-07,
"loss": 0.1186,
"step": 1037
},
{
"epoch": 0.39694072657743784,
"grad_norm": 6.277490506121325,
"learning_rate": 9.677821533141296e-07,
"loss": 0.3295,
"step": 1038
},
{
"epoch": 0.39732313575525813,
"grad_norm": 4.740829851606653,
"learning_rate": 9.679163357583416e-07,
"loss": 0.6511,
"step": 1039
},
{
"epoch": 0.3977055449330784,
"grad_norm": 3.8249444945054254,
"learning_rate": 9.680503891188934e-07,
"loss": 0.4757,
"step": 1040
},
{
"epoch": 0.39808795411089865,
"grad_norm": 4.29025279831927,
"learning_rate": 9.681843136439033e-07,
"loss": 0.4677,
"step": 1041
},
{
"epoch": 0.39847036328871893,
"grad_norm": 3.8375520435494663,
"learning_rate": 9.683181095807752e-07,
"loss": 0.4587,
"step": 1042
},
{
"epoch": 0.3988527724665392,
"grad_norm": 3.2469451926081643,
"learning_rate": 9.684517771762013e-07,
"loss": 0.3822,
"step": 1043
},
{
"epoch": 0.39923518164435945,
"grad_norm": 2.8476832091595097,
"learning_rate": 9.68585316676164e-07,
"loss": 0.4699,
"step": 1044
},
{
"epoch": 0.39961759082217974,
"grad_norm": 2.4952240146529494,
"learning_rate": 9.687187283259399e-07,
"loss": 0.3083,
"step": 1045
},
{
"epoch": 0.4,
"grad_norm": 3.6329542563068378,
"learning_rate": 9.688520123701012e-07,
"loss": 0.2969,
"step": 1046
},
{
"epoch": 0.40038240917782025,
"grad_norm": 3.894775382973128,
"learning_rate": 9.689851690525188e-07,
"loss": 0.1538,
"step": 1047
},
{
"epoch": 0.40076481835564054,
"grad_norm": 3.796723288185588,
"learning_rate": 9.69118198616366e-07,
"loss": 0.1505,
"step": 1048
},
{
"epoch": 0.4011472275334608,
"grad_norm": 3.1142443384194247,
"learning_rate": 9.692511013041197e-07,
"loss": 0.0952,
"step": 1049
},
{
"epoch": 0.40152963671128106,
"grad_norm": 5.055412323055709,
"learning_rate": 9.693838773575638e-07,
"loss": 0.3218,
"step": 1050
},
{
"epoch": 0.40191204588910134,
"grad_norm": 6.485575296006208,
"learning_rate": 9.695165270177918e-07,
"loss": 0.3999,
"step": 1051
},
{
"epoch": 0.40229445506692163,
"grad_norm": 2.793858441119319,
"learning_rate": 9.696490505252096e-07,
"loss": 0.2352,
"step": 1052
},
{
"epoch": 0.40267686424474186,
"grad_norm": 4.063080496762148,
"learning_rate": 9.69781448119537e-07,
"loss": 0.5957,
"step": 1053
},
{
"epoch": 0.40305927342256215,
"grad_norm": 1.6902282700627846,
"learning_rate": 9.699137200398118e-07,
"loss": 0.4523,
"step": 1054
},
{
"epoch": 0.40344168260038243,
"grad_norm": 3.47420780290818,
"learning_rate": 9.700458665243921e-07,
"loss": 0.4985,
"step": 1055
},
{
"epoch": 0.40382409177820267,
"grad_norm": 5.289364692022554,
"learning_rate": 9.701778878109578e-07,
"loss": 0.6845,
"step": 1056
},
{
"epoch": 0.40420650095602295,
"grad_norm": 2.7646218759762546,
"learning_rate": 9.703097841365144e-07,
"loss": 0.5403,
"step": 1057
},
{
"epoch": 0.40458891013384324,
"grad_norm": 3.1681689329670193,
"learning_rate": 9.704415557373944e-07,
"loss": 0.1867,
"step": 1058
},
{
"epoch": 0.40497131931166347,
"grad_norm": 3.386309571500823,
"learning_rate": 9.705732028492613e-07,
"loss": 0.174,
"step": 1059
},
{
"epoch": 0.40535372848948376,
"grad_norm": 4.397510274991105,
"learning_rate": 9.70704725707111e-07,
"loss": 0.2867,
"step": 1060
},
{
"epoch": 0.40573613766730404,
"grad_norm": 3.4382151466677286,
"learning_rate": 9.70836124545274e-07,
"loss": 0.1454,
"step": 1061
},
{
"epoch": 0.4061185468451243,
"grad_norm": 4.147605683289184,
"learning_rate": 9.709673995974196e-07,
"loss": 0.166,
"step": 1062
},
{
"epoch": 0.40650095602294456,
"grad_norm": 5.851581079430617,
"learning_rate": 9.710985510965567e-07,
"loss": 0.442,
"step": 1063
},
{
"epoch": 0.40688336520076485,
"grad_norm": 5.406232006947117,
"learning_rate": 9.71229579275037e-07,
"loss": 0.3363,
"step": 1064
},
{
"epoch": 0.4072657743785851,
"grad_norm": 3.3648067120705054,
"learning_rate": 9.713604843645576e-07,
"loss": 0.4594,
"step": 1065
},
{
"epoch": 0.40764818355640536,
"grad_norm": 2.5631868325853686,
"learning_rate": 9.71491266596163e-07,
"loss": 0.3384,
"step": 1066
},
{
"epoch": 0.4080305927342256,
"grad_norm": 3.177223994575539,
"learning_rate": 9.716219262002475e-07,
"loss": 0.2959,
"step": 1067
},
{
"epoch": 0.4084130019120459,
"grad_norm": 3.2975564406093048,
"learning_rate": 9.717524634065587e-07,
"loss": 0.259,
"step": 1068
},
{
"epoch": 0.40879541108986617,
"grad_norm": 3.528287201659267,
"learning_rate": 9.718828784441982e-07,
"loss": 0.3725,
"step": 1069
},
{
"epoch": 0.4091778202676864,
"grad_norm": 2.679782343039616,
"learning_rate": 9.72013171541626e-07,
"loss": 0.3988,
"step": 1070
},
{
"epoch": 0.4095602294455067,
"grad_norm": 2.4261372131515095,
"learning_rate": 9.721433429266611e-07,
"loss": 0.3528,
"step": 1071
},
{
"epoch": 0.40994263862332697,
"grad_norm": 3.072739280279131,
"learning_rate": 9.722733928264848e-07,
"loss": 0.2398,
"step": 1072
},
{
"epoch": 0.4103250478011472,
"grad_norm": 3.8742556933860697,
"learning_rate": 9.724033214676432e-07,
"loss": 0.3672,
"step": 1073
},
{
"epoch": 0.4107074569789675,
"grad_norm": 3.8825253662538373,
"learning_rate": 9.72533129076049e-07,
"loss": 0.2031,
"step": 1074
},
{
"epoch": 0.4110898661567878,
"grad_norm": 4.078706780221329,
"learning_rate": 9.72662815876984e-07,
"loss": 0.141,
"step": 1075
},
{
"epoch": 0.411472275334608,
"grad_norm": 5.947445861390411,
"learning_rate": 9.72792382095102e-07,
"loss": 0.5128,
"step": 1076
},
{
"epoch": 0.4118546845124283,
"grad_norm": 4.128468991082184,
"learning_rate": 9.729218279544307e-07,
"loss": 0.3957,
"step": 1077
},
{
"epoch": 0.4122370936902486,
"grad_norm": 3.9534924000929577,
"learning_rate": 9.73051153678374e-07,
"loss": 0.5413,
"step": 1078
},
{
"epoch": 0.4126195028680688,
"grad_norm": 2.126150605100412,
"learning_rate": 9.73180359489714e-07,
"loss": 0.3293,
"step": 1079
},
{
"epoch": 0.4130019120458891,
"grad_norm": 3.3000665195580985,
"learning_rate": 9.733094456106143e-07,
"loss": 0.5077,
"step": 1080
},
{
"epoch": 0.4133843212237094,
"grad_norm": 4.306372670944154,
"learning_rate": 9.734384122626212e-07,
"loss": 0.5619,
"step": 1081
},
{
"epoch": 0.4137667304015296,
"grad_norm": 2.64362007344908,
"learning_rate": 9.735672596666664e-07,
"loss": 0.4506,
"step": 1082
},
{
"epoch": 0.4141491395793499,
"grad_norm": 2.485214763241518,
"learning_rate": 9.736959880430697e-07,
"loss": 0.277,
"step": 1083
},
{
"epoch": 0.4145315487571702,
"grad_norm": 3.7446658025324835,
"learning_rate": 9.738245976115407e-07,
"loss": 0.2825,
"step": 1084
},
{
"epoch": 0.4149139579349904,
"grad_norm": 2.290043435163736,
"learning_rate": 9.739530885911813e-07,
"loss": 0.1555,
"step": 1085
},
{
"epoch": 0.4152963671128107,
"grad_norm": 3.027236011187704,
"learning_rate": 9.740814612004875e-07,
"loss": 0.2064,
"step": 1086
},
{
"epoch": 0.415678776290631,
"grad_norm": 3.660505117362001,
"learning_rate": 9.742097156573525e-07,
"loss": 0.1534,
"step": 1087
},
{
"epoch": 0.4160611854684512,
"grad_norm": 5.677949369209507,
"learning_rate": 9.74337852179068e-07,
"loss": 0.1978,
"step": 1088
},
{
"epoch": 0.4164435946462715,
"grad_norm": 7.4228592069466295,
"learning_rate": 9.744658709823274e-07,
"loss": 0.6633,
"step": 1089
},
{
"epoch": 0.4168260038240918,
"grad_norm": 4.292387349378489,
"learning_rate": 9.745937722832267e-07,
"loss": 0.4571,
"step": 1090
},
{
"epoch": 0.417208413001912,
"grad_norm": 2.427752641285365,
"learning_rate": 9.747215562972675e-07,
"loss": 0.3619,
"step": 1091
},
{
"epoch": 0.4175908221797323,
"grad_norm": 2.5606948262759954,
"learning_rate": 9.7484922323936e-07,
"loss": 0.635,
"step": 1092
},
{
"epoch": 0.4179732313575526,
"grad_norm": 3.664221266049891,
"learning_rate": 9.749767733238231e-07,
"loss": 0.5136,
"step": 1093
},
{
"epoch": 0.41835564053537283,
"grad_norm": 5.36261973569705,
"learning_rate": 9.751042067643886e-07,
"loss": 0.6432,
"step": 1094
},
{
"epoch": 0.4187380497131931,
"grad_norm": 2.8878135411838866,
"learning_rate": 9.752315237742018e-07,
"loss": 0.404,
"step": 1095
},
{
"epoch": 0.4191204588910134,
"grad_norm": 2.406785682401822,
"learning_rate": 9.753587245658248e-07,
"loss": 0.2421,
"step": 1096
},
{
"epoch": 0.41950286806883363,
"grad_norm": 2.4421919634113785,
"learning_rate": 9.754858093512376e-07,
"loss": 0.3162,
"step": 1097
},
{
"epoch": 0.4198852772466539,
"grad_norm": 3.9177153515856413,
"learning_rate": 9.756127783418411e-07,
"loss": 0.5863,
"step": 1098
},
{
"epoch": 0.4202676864244742,
"grad_norm": 3.9391362257114078,
"learning_rate": 9.757396317484594e-07,
"loss": 0.142,
"step": 1099
},
{
"epoch": 0.42065009560229444,
"grad_norm": 4.385574205095433,
"learning_rate": 9.758663697813404e-07,
"loss": 0.2256,
"step": 1100
},
{
"epoch": 0.4210325047801147,
"grad_norm": 5.305099538719576,
"learning_rate": 9.75992992650159e-07,
"loss": 0.6464,
"step": 1101
},
{
"epoch": 0.421414913957935,
"grad_norm": 4.7650198130080605,
"learning_rate": 9.761195005640198e-07,
"loss": 0.5413,
"step": 1102
},
{
"epoch": 0.42179732313575524,
"grad_norm": 3.604277078043537,
"learning_rate": 9.762458937314578e-07,
"loss": 0.3032,
"step": 1103
},
{
"epoch": 0.42217973231357553,
"grad_norm": 2.8768596439751475,
"learning_rate": 9.76372172360441e-07,
"loss": 0.4792,
"step": 1104
},
{
"epoch": 0.4225621414913958,
"grad_norm": 4.166289630361482,
"learning_rate": 9.76498336658373e-07,
"loss": 0.3036,
"step": 1105
},
{
"epoch": 0.42294455066921605,
"grad_norm": 3.4417821395258987,
"learning_rate": 9.76624386832094e-07,
"loss": 0.4794,
"step": 1106
},
{
"epoch": 0.42332695984703633,
"grad_norm": 6.911003306707519,
"learning_rate": 9.76750323087884e-07,
"loss": 0.2857,
"step": 1107
},
{
"epoch": 0.4237093690248566,
"grad_norm": 4.189974836396834,
"learning_rate": 9.768761456314634e-07,
"loss": 0.3152,
"step": 1108
},
{
"epoch": 0.42409177820267685,
"grad_norm": 3.0886379353453357,
"learning_rate": 9.770018546679967e-07,
"loss": 0.2533,
"step": 1109
},
{
"epoch": 0.42447418738049714,
"grad_norm": 4.58699649885338,
"learning_rate": 9.771274504020933e-07,
"loss": 0.3245,
"step": 1110
},
{
"epoch": 0.4248565965583174,
"grad_norm": 4.591641934992285,
"learning_rate": 9.772529330378098e-07,
"loss": 0.2323,
"step": 1111
},
{
"epoch": 0.42523900573613765,
"grad_norm": 5.879606672872391,
"learning_rate": 9.773783027786522e-07,
"loss": 0.1734,
"step": 1112
},
{
"epoch": 0.42562141491395794,
"grad_norm": 5.900640821408424,
"learning_rate": 9.775035598275775e-07,
"loss": 0.1883,
"step": 1113
},
{
"epoch": 0.4260038240917782,
"grad_norm": 4.3154834986532435,
"learning_rate": 9.776287043869961e-07,
"loss": 0.6946,
"step": 1114
},
{
"epoch": 0.42638623326959846,
"grad_norm": 3.7595181787484857,
"learning_rate": 9.777537366587732e-07,
"loss": 0.7462,
"step": 1115
},
{
"epoch": 0.42676864244741874,
"grad_norm": 3.384034673851337,
"learning_rate": 9.778786568442317e-07,
"loss": 0.4953,
"step": 1116
},
{
"epoch": 0.42715105162523903,
"grad_norm": 5.47317848736149,
"learning_rate": 9.780034651441531e-07,
"loss": 0.2367,
"step": 1117
},
{
"epoch": 0.42753346080305926,
"grad_norm": 3.441935781036068,
"learning_rate": 9.781281617587801e-07,
"loss": 0.3293,
"step": 1118
},
{
"epoch": 0.42791586998087955,
"grad_norm": 3.590290560639132,
"learning_rate": 9.782527468878181e-07,
"loss": 0.5157,
"step": 1119
},
{
"epoch": 0.42829827915869984,
"grad_norm": 3.6094221632216916,
"learning_rate": 9.783772207304376e-07,
"loss": 0.2926,
"step": 1120
},
{
"epoch": 0.42868068833652007,
"grad_norm": 3.451588560725235,
"learning_rate": 9.785015834852755e-07,
"loss": 0.236,
"step": 1121
},
{
"epoch": 0.42906309751434035,
"grad_norm": 3.577192293232052,
"learning_rate": 9.786258353504375e-07,
"loss": 0.1973,
"step": 1122
},
{
"epoch": 0.42944550669216064,
"grad_norm": 2.5122485514213464,
"learning_rate": 9.787499765235003e-07,
"loss": 0.264,
"step": 1123
},
{
"epoch": 0.42982791586998087,
"grad_norm": 3.255338392584085,
"learning_rate": 9.788740072015124e-07,
"loss": 0.123,
"step": 1124
},
{
"epoch": 0.43021032504780116,
"grad_norm": 4.154827362036261,
"learning_rate": 9.78997927580997e-07,
"loss": 0.1447,
"step": 1125
},
{
"epoch": 0.4305927342256214,
"grad_norm": 4.182479353795052,
"learning_rate": 9.791217378579527e-07,
"loss": 0.4986,
"step": 1126
},
{
"epoch": 0.4309751434034417,
"grad_norm": 3.758631119416443,
"learning_rate": 9.792454382278575e-07,
"loss": 0.4975,
"step": 1127
},
{
"epoch": 0.43135755258126196,
"grad_norm": 3.017336503018636,
"learning_rate": 9.79369028885668e-07,
"loss": 0.6907,
"step": 1128
},
{
"epoch": 0.4317399617590822,
"grad_norm": 2.400174583601951,
"learning_rate": 9.794925100258231e-07,
"loss": 0.2594,
"step": 1129
},
{
"epoch": 0.4321223709369025,
"grad_norm": 4.02000882617494,
"learning_rate": 9.796158818422455e-07,
"loss": 0.4644,
"step": 1130
},
{
"epoch": 0.43250478011472276,
"grad_norm": 3.448245921169389,
"learning_rate": 9.79739144528343e-07,
"loss": 0.5918,
"step": 1131
},
{
"epoch": 0.432887189292543,
"grad_norm": 3.401682459862951,
"learning_rate": 9.7986229827701e-07,
"loss": 0.4152,
"step": 1132
},
{
"epoch": 0.4332695984703633,
"grad_norm": 2.734233394770841,
"learning_rate": 9.799853432806311e-07,
"loss": 0.3426,
"step": 1133
},
{
"epoch": 0.43365200764818357,
"grad_norm": 2.4236323717916033,
"learning_rate": 9.80108279731081e-07,
"loss": 0.5017,
"step": 1134
},
{
"epoch": 0.4340344168260038,
"grad_norm": 3.5109505327608437,
"learning_rate": 9.802311078197272e-07,
"loss": 0.1354,
"step": 1135
},
{
"epoch": 0.4344168260038241,
"grad_norm": 3.139030241432602,
"learning_rate": 9.803538277374314e-07,
"loss": 0.1875,
"step": 1136
},
{
"epoch": 0.4347992351816444,
"grad_norm": 2.7639194552862896,
"learning_rate": 9.804764396745518e-07,
"loss": 0.1821,
"step": 1137
},
{
"epoch": 0.4351816443594646,
"grad_norm": 4.877676791335944,
"learning_rate": 9.805989438209442e-07,
"loss": 0.234,
"step": 1138
},
{
"epoch": 0.4355640535372849,
"grad_norm": 4.27220233592374,
"learning_rate": 9.807213403659645e-07,
"loss": 0.6155,
"step": 1139
},
{
"epoch": 0.4359464627151052,
"grad_norm": 2.3128417510023125,
"learning_rate": 9.808436294984702e-07,
"loss": 0.5823,
"step": 1140
},
{
"epoch": 0.4363288718929254,
"grad_norm": 2.6664943838097916,
"learning_rate": 9.80965811406821e-07,
"loss": 0.5333,
"step": 1141
},
{
"epoch": 0.4367112810707457,
"grad_norm": 3.5690670053012563,
"learning_rate": 9.810878862788832e-07,
"loss": 0.4513,
"step": 1142
},
{
"epoch": 0.437093690248566,
"grad_norm": 2.8073558811824375,
"learning_rate": 9.81209854302028e-07,
"loss": 0.7191,
"step": 1143
},
{
"epoch": 0.4374760994263862,
"grad_norm": 3.482803469304339,
"learning_rate": 9.813317156631366e-07,
"loss": 0.4091,
"step": 1144
},
{
"epoch": 0.4378585086042065,
"grad_norm": 3.0622469739124307,
"learning_rate": 9.814534705485993e-07,
"loss": 0.298,
"step": 1145
},
{
"epoch": 0.4382409177820268,
"grad_norm": 3.5218069020874525,
"learning_rate": 9.815751191443186e-07,
"loss": 0.1966,
"step": 1146
},
{
"epoch": 0.438623326959847,
"grad_norm": 3.3552436442245703,
"learning_rate": 9.816966616357108e-07,
"loss": 0.2007,
"step": 1147
},
{
"epoch": 0.4390057361376673,
"grad_norm": 5.2585727997837015,
"learning_rate": 9.81818098207707e-07,
"loss": 0.2871,
"step": 1148
},
{
"epoch": 0.4393881453154876,
"grad_norm": 4.389194241831785,
"learning_rate": 9.819394290447558e-07,
"loss": 0.1631,
"step": 1149
},
{
"epoch": 0.4397705544933078,
"grad_norm": 5.51060607623087,
"learning_rate": 9.820606543308238e-07,
"loss": 0.1877,
"step": 1150
},
{
"epoch": 0.4401529636711281,
"grad_norm": 5.541953555069094,
"learning_rate": 9.82181774249398e-07,
"loss": 0.5161,
"step": 1151
},
{
"epoch": 0.4405353728489484,
"grad_norm": 4.145049412413277,
"learning_rate": 9.82302788983488e-07,
"loss": 0.325,
"step": 1152
},
{
"epoch": 0.4409177820267686,
"grad_norm": 4.833186487181417,
"learning_rate": 9.824236987156263e-07,
"loss": 0.7207,
"step": 1153
},
{
"epoch": 0.4413001912045889,
"grad_norm": 2.173152630125367,
"learning_rate": 9.82544503627871e-07,
"loss": 0.5159,
"step": 1154
},
{
"epoch": 0.4416826003824092,
"grad_norm": 2.1875687724610686,
"learning_rate": 9.82665203901807e-07,
"loss": 0.5773,
"step": 1155
},
{
"epoch": 0.4420650095602294,
"grad_norm": 2.490327816874326,
"learning_rate": 9.82785799718548e-07,
"loss": 0.5218,
"step": 1156
},
{
"epoch": 0.4424474187380497,
"grad_norm": 2.9280410412585467,
"learning_rate": 9.829062912587372e-07,
"loss": 0.3466,
"step": 1157
},
{
"epoch": 0.44282982791587,
"grad_norm": 2.2559649554229417,
"learning_rate": 9.830266787025507e-07,
"loss": 0.495,
"step": 1158
},
{
"epoch": 0.44321223709369023,
"grad_norm": 2.5825451499633303,
"learning_rate": 9.83146962229697e-07,
"loss": 0.2198,
"step": 1159
},
{
"epoch": 0.4435946462715105,
"grad_norm": 2.5606514380526155,
"learning_rate": 9.832671420194203e-07,
"loss": 0.126,
"step": 1160
},
{
"epoch": 0.4439770554493308,
"grad_norm": 3.403153017767686,
"learning_rate": 9.83387218250501e-07,
"loss": 0.1227,
"step": 1161
},
{
"epoch": 0.44435946462715104,
"grad_norm": 3.6106131935818038,
"learning_rate": 9.835071911012583e-07,
"loss": 0.1078,
"step": 1162
},
{
"epoch": 0.4447418738049713,
"grad_norm": 5.662654728070213,
"learning_rate": 9.836270607495506e-07,
"loss": 0.3093,
"step": 1163
},
{
"epoch": 0.4451242829827916,
"grad_norm": 4.739358928803586,
"learning_rate": 9.837468273727776e-07,
"loss": 0.3841,
"step": 1164
},
{
"epoch": 0.44550669216061184,
"grad_norm": 4.695340326261659,
"learning_rate": 9.83866491147883e-07,
"loss": 0.678,
"step": 1165
},
{
"epoch": 0.4458891013384321,
"grad_norm": 3.343741443658737,
"learning_rate": 9.83986052251354e-07,
"loss": 0.5099,
"step": 1166
},
{
"epoch": 0.4462715105162524,
"grad_norm": 4.670818707014738,
"learning_rate": 9.841055108592244e-07,
"loss": 0.4155,
"step": 1167
},
{
"epoch": 0.44665391969407264,
"grad_norm": 3.5820297856884276,
"learning_rate": 9.842248671470756e-07,
"loss": 0.5674,
"step": 1168
},
{
"epoch": 0.44703632887189293,
"grad_norm": 3.620177909315062,
"learning_rate": 9.84344121290038e-07,
"loss": 0.5238,
"step": 1169
},
{
"epoch": 0.4474187380497132,
"grad_norm": 2.889158629597049,
"learning_rate": 9.84463273462793e-07,
"loss": 0.1835,
"step": 1170
},
{
"epoch": 0.44780114722753345,
"grad_norm": 3.3970537547478528,
"learning_rate": 9.845823238395744e-07,
"loss": 0.4899,
"step": 1171
},
{
"epoch": 0.44818355640535373,
"grad_norm": 3.7631936989471986,
"learning_rate": 9.847012725941693e-07,
"loss": 0.3491,
"step": 1172
},
{
"epoch": 0.448565965583174,
"grad_norm": 4.2082091712890515,
"learning_rate": 9.84820119899921e-07,
"loss": 0.2193,
"step": 1173
},
{
"epoch": 0.44894837476099425,
"grad_norm": 3.8214286936618596,
"learning_rate": 9.849388659297287e-07,
"loss": 0.1701,
"step": 1174
},
{
"epoch": 0.44933078393881454,
"grad_norm": 3.445321634501071,
"learning_rate": 9.850575108560506e-07,
"loss": 0.1214,
"step": 1175
},
{
"epoch": 0.4497131931166348,
"grad_norm": 7.539568726932249,
"learning_rate": 9.851760548509042e-07,
"loss": 0.6822,
"step": 1176
},
{
"epoch": 0.45009560229445505,
"grad_norm": 4.843640726148957,
"learning_rate": 9.85294498085869e-07,
"loss": 0.7945,
"step": 1177
},
{
"epoch": 0.45047801147227534,
"grad_norm": 3.354664041040037,
"learning_rate": 9.854128407320875e-07,
"loss": 0.5656,
"step": 1178
},
{
"epoch": 0.45086042065009563,
"grad_norm": 3.114434960615066,
"learning_rate": 9.855310829602656e-07,
"loss": 0.5782,
"step": 1179
},
{
"epoch": 0.45124282982791586,
"grad_norm": 2.023679976497725,
"learning_rate": 9.856492249406758e-07,
"loss": 0.4287,
"step": 1180
},
{
"epoch": 0.45162523900573615,
"grad_norm": 2.9935946430201184,
"learning_rate": 9.857672668431576e-07,
"loss": 0.4449,
"step": 1181
},
{
"epoch": 0.45200764818355643,
"grad_norm": 3.0199307551245393,
"learning_rate": 9.85885208837119e-07,
"loss": 0.2413,
"step": 1182
},
{
"epoch": 0.45239005736137666,
"grad_norm": 3.2386902848763635,
"learning_rate": 9.860030510915388e-07,
"loss": 0.2867,
"step": 1183
},
{
"epoch": 0.45277246653919695,
"grad_norm": 1.7672623325899193,
"learning_rate": 9.86120793774967e-07,
"loss": 0.3667,
"step": 1184
},
{
"epoch": 0.45315487571701724,
"grad_norm": 2.526398707585068,
"learning_rate": 9.86238437055527e-07,
"loss": 0.2138,
"step": 1185
},
{
"epoch": 0.45353728489483747,
"grad_norm": 3.4035631134247284,
"learning_rate": 9.863559811009169e-07,
"loss": 0.1627,
"step": 1186
},
{
"epoch": 0.45391969407265775,
"grad_norm": 4.542402132079084,
"learning_rate": 9.864734260784098e-07,
"loss": 0.1862,
"step": 1187
},
{
"epoch": 0.454302103250478,
"grad_norm": 6.694197393829867,
"learning_rate": 9.865907721548575e-07,
"loss": 0.2665,
"step": 1188
},
{
"epoch": 0.45468451242829827,
"grad_norm": 4.0984265355302725,
"learning_rate": 9.8670801949669e-07,
"loss": 0.5322,
"step": 1189
},
{
"epoch": 0.45506692160611856,
"grad_norm": 4.028227863522493,
"learning_rate": 9.868251682699173e-07,
"loss": 0.5913,
"step": 1190
},
{
"epoch": 0.4554493307839388,
"grad_norm": 2.6755965343120414,
"learning_rate": 9.869422186401317e-07,
"loss": 0.2668,
"step": 1191
},
{
"epoch": 0.4558317399617591,
"grad_norm": 3.1639273155438374,
"learning_rate": 9.87059170772508e-07,
"loss": 0.3166,
"step": 1192
},
{
"epoch": 0.45621414913957936,
"grad_norm": 2.9884248586816313,
"learning_rate": 9.87176024831806e-07,
"loss": 0.3124,
"step": 1193
},
{
"epoch": 0.4565965583173996,
"grad_norm": 2.3033321209843156,
"learning_rate": 9.872927809823707e-07,
"loss": 0.1689,
"step": 1194
},
{
"epoch": 0.4569789674952199,
"grad_norm": 3.2703745493900755,
"learning_rate": 9.874094393881347e-07,
"loss": 0.2984,
"step": 1195
},
{
"epoch": 0.45736137667304017,
"grad_norm": 2.8477148565428396,
"learning_rate": 9.875260002126198e-07,
"loss": 0.2651,
"step": 1196
},
{
"epoch": 0.4577437858508604,
"grad_norm": 3.408230686896076,
"learning_rate": 9.876424636189368e-07,
"loss": 0.3628,
"step": 1197
},
{
"epoch": 0.4581261950286807,
"grad_norm": 3.6999295769252885,
"learning_rate": 9.877588297697883e-07,
"loss": 0.2163,
"step": 1198
},
{
"epoch": 0.45850860420650097,
"grad_norm": 2.585867705080345,
"learning_rate": 9.878750988274699e-07,
"loss": 0.1576,
"step": 1199
},
{
"epoch": 0.4588910133843212,
"grad_norm": 3.341464108030177,
"learning_rate": 9.879912709538705e-07,
"loss": 0.1122,
"step": 1200
},
{
"epoch": 0.4592734225621415,
"grad_norm": 4.997080522787351,
"learning_rate": 9.881073463104754e-07,
"loss": 0.5105,
"step": 1201
},
{
"epoch": 0.4596558317399618,
"grad_norm": 4.627752361454242,
"learning_rate": 9.882233250583664e-07,
"loss": 0.5844,
"step": 1202
},
{
"epoch": 0.460038240917782,
"grad_norm": 2.9673988881677205,
"learning_rate": 9.883392073582227e-07,
"loss": 0.4366,
"step": 1203
},
{
"epoch": 0.4604206500956023,
"grad_norm": 1.9153244529546578,
"learning_rate": 9.884549933703243e-07,
"loss": 0.3027,
"step": 1204
},
{
"epoch": 0.4608030592734226,
"grad_norm": 4.689452921634211,
"learning_rate": 9.885706832545508e-07,
"loss": 0.6043,
"step": 1205
},
{
"epoch": 0.4611854684512428,
"grad_norm": 3.31635650130163,
"learning_rate": 9.886862771703845e-07,
"loss": 0.4529,
"step": 1206
},
{
"epoch": 0.4615678776290631,
"grad_norm": 3.965577787654834,
"learning_rate": 9.88801775276911e-07,
"loss": 0.6734,
"step": 1207
},
{
"epoch": 0.4619502868068834,
"grad_norm": 2.7480814049780675,
"learning_rate": 9.88917177732821e-07,
"loss": 0.3801,
"step": 1208
},
{
"epoch": 0.4623326959847036,
"grad_norm": 4.818329092362784,
"learning_rate": 9.890324846964104e-07,
"loss": 0.3093,
"step": 1209
},
{
"epoch": 0.4627151051625239,
"grad_norm": 3.9353637955632736,
"learning_rate": 9.891476963255836e-07,
"loss": 0.1566,
"step": 1210
},
{
"epoch": 0.4630975143403442,
"grad_norm": 3.958589779761676,
"learning_rate": 9.892628127778526e-07,
"loss": 0.3374,
"step": 1211
},
{
"epoch": 0.4634799235181644,
"grad_norm": 5.670833775812039,
"learning_rate": 9.8937783421034e-07,
"loss": 0.219,
"step": 1212
},
{
"epoch": 0.4638623326959847,
"grad_norm": 10.89393948397313,
"learning_rate": 9.894927607797798e-07,
"loss": 0.3929,
"step": 1213
},
{
"epoch": 0.464244741873805,
"grad_norm": 3.952266675322374,
"learning_rate": 9.896075926425177e-07,
"loss": 0.4048,
"step": 1214
},
{
"epoch": 0.4646271510516252,
"grad_norm": 4.457437561932895,
"learning_rate": 9.89722329954514e-07,
"loss": 0.6058,
"step": 1215
},
{
"epoch": 0.4650095602294455,
"grad_norm": 2.677637322445704,
"learning_rate": 9.898369728713437e-07,
"loss": 0.345,
"step": 1216
},
{
"epoch": 0.4653919694072658,
"grad_norm": 3.192019680278025,
"learning_rate": 9.899515215481983e-07,
"loss": 0.3367,
"step": 1217
},
{
"epoch": 0.465774378585086,
"grad_norm": 3.4701386842304975,
"learning_rate": 9.900659761398872e-07,
"loss": 0.5685,
"step": 1218
},
{
"epoch": 0.4661567877629063,
"grad_norm": 4.125531480246737,
"learning_rate": 9.901803368008373e-07,
"loss": 0.5088,
"step": 1219
},
{
"epoch": 0.4665391969407266,
"grad_norm": 4.025808757774617,
"learning_rate": 9.902946036850975e-07,
"loss": 0.4209,
"step": 1220
},
{
"epoch": 0.46692160611854683,
"grad_norm": 3.4272727720703533,
"learning_rate": 9.904087769463365e-07,
"loss": 0.286,
"step": 1221
},
{
"epoch": 0.4673040152963671,
"grad_norm": 2.8436544213634063,
"learning_rate": 9.905228567378466e-07,
"loss": 0.2031,
"step": 1222
},
{
"epoch": 0.4676864244741874,
"grad_norm": 4.01348321104643,
"learning_rate": 9.906368432125435e-07,
"loss": 0.264,
"step": 1223
},
{
"epoch": 0.46806883365200763,
"grad_norm": 5.429052170282707,
"learning_rate": 9.907507365229678e-07,
"loss": 0.1528,
"step": 1224
},
{
"epoch": 0.4684512428298279,
"grad_norm": 5.180266887314177,
"learning_rate": 9.908645368212868e-07,
"loss": 0.1564,
"step": 1225
},
{
"epoch": 0.4688336520076482,
"grad_norm": 5.246787405276859,
"learning_rate": 9.909782442592948e-07,
"loss": 0.5546,
"step": 1226
},
{
"epoch": 0.46921606118546844,
"grad_norm": 3.9574744380238647,
"learning_rate": 9.910918589884153e-07,
"loss": 0.5016,
"step": 1227
},
{
"epoch": 0.4695984703632887,
"grad_norm": 3.5726202720433826,
"learning_rate": 9.912053811597014e-07,
"loss": 0.2558,
"step": 1228
},
{
"epoch": 0.469980879541109,
"grad_norm": 2.7987347605240522,
"learning_rate": 9.913188109238372e-07,
"loss": 0.2549,
"step": 1229
},
{
"epoch": 0.47036328871892924,
"grad_norm": 3.427817284431774,
"learning_rate": 9.9143214843114e-07,
"loss": 0.5589,
"step": 1230
},
{
"epoch": 0.4707456978967495,
"grad_norm": 3.163584129057733,
"learning_rate": 9.915453938315595e-07,
"loss": 0.2749,
"step": 1231
},
{
"epoch": 0.4711281070745698,
"grad_norm": 3.300248431477948,
"learning_rate": 9.916585472746808e-07,
"loss": 0.3889,
"step": 1232
},
{
"epoch": 0.47151051625239004,
"grad_norm": 2.516873446763163,
"learning_rate": 9.917716089097243e-07,
"loss": 0.3703,
"step": 1233
},
{
"epoch": 0.47189292543021033,
"grad_norm": 16.50220821516207,
"learning_rate": 9.918845788855485e-07,
"loss": 0.1553,
"step": 1234
},
{
"epoch": 0.4722753346080306,
"grad_norm": 3.361109653889434,
"learning_rate": 9.91997457350649e-07,
"loss": 0.2914,
"step": 1235
},
{
"epoch": 0.47265774378585085,
"grad_norm": 3.522181386406725,
"learning_rate": 9.921102444531612e-07,
"loss": 0.1333,
"step": 1236
},
{
"epoch": 0.47304015296367113,
"grad_norm": 5.504639489394662,
"learning_rate": 9.92222940340862e-07,
"loss": 0.1475,
"step": 1237
},
{
"epoch": 0.4734225621414914,
"grad_norm": 5.981770479044214,
"learning_rate": 9.923355451611689e-07,
"loss": 0.305,
"step": 1238
},
{
"epoch": 0.47380497131931165,
"grad_norm": 4.278875207431444,
"learning_rate": 9.924480590611424e-07,
"loss": 0.7047,
"step": 1239
},
{
"epoch": 0.47418738049713194,
"grad_norm": 3.7687764889622843,
"learning_rate": 9.92560482187488e-07,
"loss": 0.3426,
"step": 1240
},
{
"epoch": 0.4745697896749522,
"grad_norm": 2.821927862844503,
"learning_rate": 9.926728146865552e-07,
"loss": 0.5761,
"step": 1241
},
{
"epoch": 0.47495219885277246,
"grad_norm": 2.7631673291650007,
"learning_rate": 9.927850567043408e-07,
"loss": 0.4746,
"step": 1242
},
{
"epoch": 0.47533460803059274,
"grad_norm": 2.434910499134197,
"learning_rate": 9.928972083864887e-07,
"loss": 0.4313,
"step": 1243
},
{
"epoch": 0.47571701720841303,
"grad_norm": 2.7586319438112827,
"learning_rate": 9.930092698782916e-07,
"loss": 0.4159,
"step": 1244
},
{
"epoch": 0.47609942638623326,
"grad_norm": 3.0803940828055265,
"learning_rate": 9.931212413246912e-07,
"loss": 0.2955,
"step": 1245
},
{
"epoch": 0.47648183556405355,
"grad_norm": 3.7571820465812333,
"learning_rate": 9.932331228702814e-07,
"loss": 0.2781,
"step": 1246
},
{
"epoch": 0.4768642447418738,
"grad_norm": 4.616070762396008,
"learning_rate": 9.933449146593071e-07,
"loss": 0.3388,
"step": 1247
},
{
"epoch": 0.47724665391969406,
"grad_norm": 2.935917529514803,
"learning_rate": 9.934566168356667e-07,
"loss": 0.1672,
"step": 1248
},
{
"epoch": 0.47762906309751435,
"grad_norm": 5.0556371819656745,
"learning_rate": 9.935682295429127e-07,
"loss": 0.1328,
"step": 1249
},
{
"epoch": 0.4780114722753346,
"grad_norm": 4.746674427569042,
"learning_rate": 9.93679752924253e-07,
"loss": 0.1172,
"step": 1250
},
{
"epoch": 0.47839388145315487,
"grad_norm": 5.159086429836909,
"learning_rate": 9.93791187122552e-07,
"loss": 0.3563,
"step": 1251
},
{
"epoch": 0.47877629063097515,
"grad_norm": 3.385802042618818,
"learning_rate": 9.939025322803313e-07,
"loss": 0.5295,
"step": 1252
},
{
"epoch": 0.4791586998087954,
"grad_norm": 3.145228414079556,
"learning_rate": 9.940137885397717e-07,
"loss": 0.3932,
"step": 1253
},
{
"epoch": 0.47954110898661567,
"grad_norm": 2.2029955670367993,
"learning_rate": 9.941249560427134e-07,
"loss": 0.2603,
"step": 1254
},
{
"epoch": 0.47992351816443596,
"grad_norm": 2.91587793493996,
"learning_rate": 9.942360349306571e-07,
"loss": 0.4738,
"step": 1255
},
{
"epoch": 0.4803059273422562,
"grad_norm": 4.883861996587228,
"learning_rate": 9.94347025344766e-07,
"loss": 0.5368,
"step": 1256
},
{
"epoch": 0.4806883365200765,
"grad_norm": 2.9874481835121296,
"learning_rate": 9.94457927425866e-07,
"loss": 0.2862,
"step": 1257
},
{
"epoch": 0.48107074569789676,
"grad_norm": 3.496120482614276,
"learning_rate": 9.945687413144468e-07,
"loss": 0.3032,
"step": 1258
},
{
"epoch": 0.481453154875717,
"grad_norm": 2.92530042621577,
"learning_rate": 9.946794671506636e-07,
"loss": 0.4655,
"step": 1259
},
{
"epoch": 0.4818355640535373,
"grad_norm": 3.2392137977017943,
"learning_rate": 9.947901050743373e-07,
"loss": 0.2499,
"step": 1260
},
{
"epoch": 0.48221797323135757,
"grad_norm": 5.070935603384918,
"learning_rate": 9.949006552249564e-07,
"loss": 0.1595,
"step": 1261
},
{
"epoch": 0.4826003824091778,
"grad_norm": 3.7492990474444645,
"learning_rate": 9.950111177416775e-07,
"loss": 0.1476,
"step": 1262
},
{
"epoch": 0.4829827915869981,
"grad_norm": 4.9714620614137415,
"learning_rate": 9.951214927633269e-07,
"loss": 0.3147,
"step": 1263
},
{
"epoch": 0.48336520076481837,
"grad_norm": 3.678668940147185,
"learning_rate": 9.952317804284008e-07,
"loss": 0.3862,
"step": 1264
},
{
"epoch": 0.4837476099426386,
"grad_norm": 4.033745045311636,
"learning_rate": 9.95341980875067e-07,
"loss": 0.4374,
"step": 1265
},
{
"epoch": 0.4841300191204589,
"grad_norm": 2.5253120758279097,
"learning_rate": 9.954520942411654e-07,
"loss": 0.3187,
"step": 1266
},
{
"epoch": 0.4845124282982792,
"grad_norm": 2.629254027626614,
"learning_rate": 9.955621206642103e-07,
"loss": 0.4113,
"step": 1267
},
{
"epoch": 0.4848948374760994,
"grad_norm": 3.4575327364019297,
"learning_rate": 9.956720602813898e-07,
"loss": 0.3076,
"step": 1268
},
{
"epoch": 0.4852772466539197,
"grad_norm": 1.6920858167705082,
"learning_rate": 9.957819132295675e-07,
"loss": 0.3438,
"step": 1269
},
{
"epoch": 0.48565965583174,
"grad_norm": 2.9171764109992773,
"learning_rate": 9.958916796452842e-07,
"loss": 0.3557,
"step": 1270
},
{
"epoch": 0.4860420650095602,
"grad_norm": 3.131474311393124,
"learning_rate": 9.960013596647574e-07,
"loss": 0.203,
"step": 1271
},
{
"epoch": 0.4864244741873805,
"grad_norm": 2.664857896662694,
"learning_rate": 9.961109534238842e-07,
"loss": 0.2681,
"step": 1272
},
{
"epoch": 0.4868068833652008,
"grad_norm": 3.2868231537134185,
"learning_rate": 9.962204610582404e-07,
"loss": 0.1471,
"step": 1273
},
{
"epoch": 0.487189292543021,
"grad_norm": 5.023350550253528,
"learning_rate": 9.96329882703083e-07,
"loss": 0.1718,
"step": 1274
},
{
"epoch": 0.4875717017208413,
"grad_norm": 4.181604591011255,
"learning_rate": 9.964392184933502e-07,
"loss": 0.1601,
"step": 1275
},
{
"epoch": 0.4879541108986616,
"grad_norm": 4.9286020961146955,
"learning_rate": 9.965484685636635e-07,
"loss": 0.3611,
"step": 1276
},
{
"epoch": 0.4883365200764818,
"grad_norm": 3.785761892014254,
"learning_rate": 9.966576330483273e-07,
"loss": 0.5685,
"step": 1277
},
{
"epoch": 0.4887189292543021,
"grad_norm": 3.2837802395628066,
"learning_rate": 9.96766712081331e-07,
"loss": 0.5583,
"step": 1278
},
{
"epoch": 0.4891013384321224,
"grad_norm": 3.5432906672246642,
"learning_rate": 9.968757057963495e-07,
"loss": 0.4116,
"step": 1279
},
{
"epoch": 0.4894837476099426,
"grad_norm": 3.013914898253767,
"learning_rate": 9.969846143267443e-07,
"loss": 0.4147,
"step": 1280
},
{
"epoch": 0.4898661567877629,
"grad_norm": 4.501812252066544,
"learning_rate": 9.97093437805564e-07,
"loss": 0.5485,
"step": 1281
},
{
"epoch": 0.4902485659655832,
"grad_norm": 2.827393157856154,
"learning_rate": 9.972021763655468e-07,
"loss": 0.3098,
"step": 1282
},
{
"epoch": 0.4906309751434034,
"grad_norm": 3.0924155700062674,
"learning_rate": 9.973108301391195e-07,
"loss": 0.2388,
"step": 1283
},
{
"epoch": 0.4910133843212237,
"grad_norm": 2.9308757355975787,
"learning_rate": 9.974193992583994e-07,
"loss": 0.1811,
"step": 1284
},
{
"epoch": 0.491395793499044,
"grad_norm": 3.1683061939771453,
"learning_rate": 9.975278838551958e-07,
"loss": 0.259,
"step": 1285
},
{
"epoch": 0.49177820267686423,
"grad_norm": 5.305806869138074,
"learning_rate": 9.976362840610098e-07,
"loss": 0.1055,
"step": 1286
},
{
"epoch": 0.4921606118546845,
"grad_norm": 5.969263823596937,
"learning_rate": 9.977446000070363e-07,
"loss": 0.2077,
"step": 1287
},
{
"epoch": 0.4925430210325048,
"grad_norm": 4.696538527080613,
"learning_rate": 9.978528318241642e-07,
"loss": 0.3235,
"step": 1288
},
{
"epoch": 0.49292543021032503,
"grad_norm": 3.6845788538530737,
"learning_rate": 9.979609796429774e-07,
"loss": 0.4144,
"step": 1289
},
{
"epoch": 0.4933078393881453,
"grad_norm": 4.982568108398919,
"learning_rate": 9.980690435937572e-07,
"loss": 0.6038,
"step": 1290
},
{
"epoch": 0.4936902485659656,
"grad_norm": 2.4051702479577983,
"learning_rate": 9.981770238064806e-07,
"loss": 0.4925,
"step": 1291
},
{
"epoch": 0.49407265774378584,
"grad_norm": 2.7142936753034363,
"learning_rate": 9.982849204108237e-07,
"loss": 0.4256,
"step": 1292
},
{
"epoch": 0.4944550669216061,
"grad_norm": 3.632771415888318,
"learning_rate": 9.983927335361607e-07,
"loss": 0.7238,
"step": 1293
},
{
"epoch": 0.4948374760994264,
"grad_norm": 2.7053264848194076,
"learning_rate": 9.985004633115662e-07,
"loss": 0.3043,
"step": 1294
},
{
"epoch": 0.49521988527724664,
"grad_norm": 2.6085726480924345,
"learning_rate": 9.986081098658163e-07,
"loss": 0.4205,
"step": 1295
},
{
"epoch": 0.4956022944550669,
"grad_norm": 3.171201136373861,
"learning_rate": 9.987156733273876e-07,
"loss": 0.2772,
"step": 1296
},
{
"epoch": 0.4959847036328872,
"grad_norm": 3.872751875899766,
"learning_rate": 9.988231538244609e-07,
"loss": 0.2868,
"step": 1297
},
{
"epoch": 0.49636711281070744,
"grad_norm": 2.6882358696856294,
"learning_rate": 9.989305514849188e-07,
"loss": 0.1262,
"step": 1298
},
{
"epoch": 0.49674952198852773,
"grad_norm": 4.098589649932069,
"learning_rate": 9.990378664363505e-07,
"loss": 0.1473,
"step": 1299
},
{
"epoch": 0.497131931166348,
"grad_norm": 3.4145372003988643,
"learning_rate": 9.991450988060493e-07,
"loss": 0.1244,
"step": 1300
},
{
"epoch": 0.49751434034416825,
"grad_norm": 4.759577080988201,
"learning_rate": 9.992522487210147e-07,
"loss": 0.3778,
"step": 1301
},
{
"epoch": 0.49789674952198854,
"grad_norm": 3.4069575703506247,
"learning_rate": 9.993593163079546e-07,
"loss": 0.2946,
"step": 1302
},
{
"epoch": 0.4982791586998088,
"grad_norm": 2.3063970916899077,
"learning_rate": 9.994663016932842e-07,
"loss": 0.4804,
"step": 1303
},
{
"epoch": 0.49866156787762905,
"grad_norm": 2.736609543154999,
"learning_rate": 9.99573205003128e-07,
"loss": 0.2697,
"step": 1304
},
{
"epoch": 0.49904397705544934,
"grad_norm": 3.491110548163302,
"learning_rate": 9.9968002636332e-07,
"loss": 0.4484,
"step": 1305
},
{
"epoch": 0.4994263862332696,
"grad_norm": 3.0667836339730554,
"learning_rate": 9.99786765899406e-07,
"loss": 0.2802,
"step": 1306
},
{
"epoch": 0.49980879541108986,
"grad_norm": 2.063057217320862,
"learning_rate": 9.998934237366427e-07,
"loss": 0.4247,
"step": 1307
},
{
"epoch": 0.5001912045889101,
"grad_norm": 2.096118327710105,
"learning_rate": 1e-06,
"loss": 0.3153,
"step": 1308
},
{
"epoch": 0.5005736137667304,
"grad_norm": 3.693911264581626,
"learning_rate": 1e-06,
"loss": 0.2241,
"step": 1309
},
{
"epoch": 0.5009560229445507,
"grad_norm": 3.4515142351940358,
"learning_rate": 1e-06,
"loss": 0.4212,
"step": 1310
},
{
"epoch": 0.501338432122371,
"grad_norm": 4.489151912533015,
"learning_rate": 1e-06,
"loss": 0.1585,
"step": 1311
},
{
"epoch": 0.5017208413001912,
"grad_norm": 4.848606751094361,
"learning_rate": 1e-06,
"loss": 0.1356,
"step": 1312
},
{
"epoch": 0.5021032504780115,
"grad_norm": 4.52506228240761,
"learning_rate": 1e-06,
"loss": 0.1422,
"step": 1313
},
{
"epoch": 0.5024856596558317,
"grad_norm": 3.254764146126662,
"learning_rate": 1e-06,
"loss": 0.5845,
"step": 1314
},
{
"epoch": 0.502868068833652,
"grad_norm": 4.1668970934412455,
"learning_rate": 1e-06,
"loss": 0.5553,
"step": 1315
},
{
"epoch": 0.5032504780114723,
"grad_norm": 3.452117634896171,
"learning_rate": 1e-06,
"loss": 0.5817,
"step": 1316
},
{
"epoch": 0.5036328871892926,
"grad_norm": 3.2847363070902205,
"learning_rate": 1e-06,
"loss": 0.3663,
"step": 1317
},
{
"epoch": 0.5040152963671128,
"grad_norm": 3.2185864453288633,
"learning_rate": 1e-06,
"loss": 0.3871,
"step": 1318
},
{
"epoch": 0.5043977055449331,
"grad_norm": 2.2661042857469327,
"learning_rate": 1e-06,
"loss": 0.3103,
"step": 1319
},
{
"epoch": 0.5047801147227533,
"grad_norm": 3.6130270708739736,
"learning_rate": 1e-06,
"loss": 0.2834,
"step": 1320
},
{
"epoch": 0.5051625239005736,
"grad_norm": 3.068598067179928,
"learning_rate": 1e-06,
"loss": 0.1962,
"step": 1321
},
{
"epoch": 0.5055449330783939,
"grad_norm": 2.531267896047366,
"learning_rate": 1e-06,
"loss": 0.2529,
"step": 1322
},
{
"epoch": 0.5059273422562142,
"grad_norm": 5.057852790207729,
"learning_rate": 1e-06,
"loss": 0.1403,
"step": 1323
},
{
"epoch": 0.5063097514340344,
"grad_norm": 4.646924402504079,
"learning_rate": 1e-06,
"loss": 0.0901,
"step": 1324
},
{
"epoch": 0.5066921606118547,
"grad_norm": 3.614905937522274,
"learning_rate": 1e-06,
"loss": 0.1286,
"step": 1325
},
{
"epoch": 0.5070745697896749,
"grad_norm": 5.897274564579732,
"learning_rate": 1e-06,
"loss": 0.3708,
"step": 1326
},
{
"epoch": 0.5074569789674952,
"grad_norm": 3.4151865071294427,
"learning_rate": 1e-06,
"loss": 0.45,
"step": 1327
},
{
"epoch": 0.5078393881453155,
"grad_norm": 2.4104035434568982,
"learning_rate": 1e-06,
"loss": 0.3933,
"step": 1328
},
{
"epoch": 0.5082217973231358,
"grad_norm": 2.700161406319671,
"learning_rate": 1e-06,
"loss": 0.3618,
"step": 1329
},
{
"epoch": 0.5086042065009561,
"grad_norm": 3.2288464459211887,
"learning_rate": 1e-06,
"loss": 0.3783,
"step": 1330
},
{
"epoch": 0.5089866156787763,
"grad_norm": 2.586615804814528,
"learning_rate": 1e-06,
"loss": 0.1344,
"step": 1331
},
{
"epoch": 0.5093690248565965,
"grad_norm": 3.315912180695968,
"learning_rate": 1e-06,
"loss": 0.3309,
"step": 1332
},
{
"epoch": 0.5097514340344168,
"grad_norm": 3.367855704933834,
"learning_rate": 1e-06,
"loss": 0.2817,
"step": 1333
},
{
"epoch": 0.5101338432122371,
"grad_norm": 3.674560084579588,
"learning_rate": 1e-06,
"loss": 0.1843,
"step": 1334
},
{
"epoch": 0.5105162523900574,
"grad_norm": 4.558705881761429,
"learning_rate": 1e-06,
"loss": 0.2367,
"step": 1335
},
{
"epoch": 0.5108986615678777,
"grad_norm": 4.22902800735144,
"learning_rate": 1e-06,
"loss": 0.1465,
"step": 1336
},
{
"epoch": 0.511281070745698,
"grad_norm": 5.283314318377014,
"learning_rate": 1e-06,
"loss": 0.1821,
"step": 1337
},
{
"epoch": 0.5116634799235181,
"grad_norm": 9.455108182339645,
"learning_rate": 1e-06,
"loss": 0.231,
"step": 1338
},
{
"epoch": 0.5120458891013384,
"grad_norm": 4.439166118912293,
"learning_rate": 1e-06,
"loss": 0.5254,
"step": 1339
},
{
"epoch": 0.5124282982791587,
"grad_norm": 3.2336246523102847,
"learning_rate": 1e-06,
"loss": 0.7046,
"step": 1340
},
{
"epoch": 0.512810707456979,
"grad_norm": 2.6178188188833382,
"learning_rate": 1e-06,
"loss": 0.5535,
"step": 1341
},
{
"epoch": 0.5131931166347993,
"grad_norm": 2.8595616013971545,
"learning_rate": 1e-06,
"loss": 0.2644,
"step": 1342
},
{
"epoch": 0.5135755258126194,
"grad_norm": 2.3391660495381705,
"learning_rate": 1e-06,
"loss": 0.2714,
"step": 1343
},
{
"epoch": 0.5139579349904397,
"grad_norm": 2.3435370793584385,
"learning_rate": 1e-06,
"loss": 0.2043,
"step": 1344
},
{
"epoch": 0.51434034416826,
"grad_norm": 2.6533094398183104,
"learning_rate": 1e-06,
"loss": 0.1981,
"step": 1345
},
{
"epoch": 0.5147227533460803,
"grad_norm": 2.4914698030745965,
"learning_rate": 1e-06,
"loss": 0.3858,
"step": 1346
},
{
"epoch": 0.5151051625239006,
"grad_norm": 2.8747949941769626,
"learning_rate": 1e-06,
"loss": 0.2616,
"step": 1347
},
{
"epoch": 0.5154875717017209,
"grad_norm": 3.3837906125988515,
"learning_rate": 1e-06,
"loss": 0.1699,
"step": 1348
},
{
"epoch": 0.5158699808795411,
"grad_norm": 3.262904151642769,
"learning_rate": 1e-06,
"loss": 0.1372,
"step": 1349
},
{
"epoch": 0.5162523900573613,
"grad_norm": 2.9114792242420964,
"learning_rate": 1e-06,
"loss": 0.075,
"step": 1350
},
{
"epoch": 0.5166347992351816,
"grad_norm": 6.668526517203372,
"learning_rate": 1e-06,
"loss": 0.3453,
"step": 1351
},
{
"epoch": 0.5170172084130019,
"grad_norm": 5.45561340743341,
"learning_rate": 1e-06,
"loss": 0.7369,
"step": 1352
},
{
"epoch": 0.5173996175908222,
"grad_norm": 2.50805814971923,
"learning_rate": 1e-06,
"loss": 0.4992,
"step": 1353
},
{
"epoch": 0.5177820267686425,
"grad_norm": 2.969272125909418,
"learning_rate": 1e-06,
"loss": 0.4825,
"step": 1354
},
{
"epoch": 0.5181644359464627,
"grad_norm": 2.791645467497392,
"learning_rate": 1e-06,
"loss": 0.5708,
"step": 1355
},
{
"epoch": 0.518546845124283,
"grad_norm": 4.315351345216982,
"learning_rate": 1e-06,
"loss": 0.3612,
"step": 1356
},
{
"epoch": 0.5189292543021032,
"grad_norm": 2.652557679033688,
"learning_rate": 1e-06,
"loss": 0.2474,
"step": 1357
},
{
"epoch": 0.5193116634799235,
"grad_norm": 3.283614405753792,
"learning_rate": 1e-06,
"loss": 0.2597,
"step": 1358
},
{
"epoch": 0.5196940726577438,
"grad_norm": 3.046856219894045,
"learning_rate": 1e-06,
"loss": 0.1428,
"step": 1359
},
{
"epoch": 0.5200764818355641,
"grad_norm": 4.348918227812207,
"learning_rate": 1e-06,
"loss": 0.2275,
"step": 1360
},
{
"epoch": 0.5204588910133843,
"grad_norm": 3.8700391947809294,
"learning_rate": 1e-06,
"loss": 0.1018,
"step": 1361
},
{
"epoch": 0.5208413001912046,
"grad_norm": 3.9402543379897286,
"learning_rate": 1e-06,
"loss": 0.1486,
"step": 1362
},
{
"epoch": 0.5212237093690248,
"grad_norm": 5.106967370431914,
"learning_rate": 1e-06,
"loss": 0.2618,
"step": 1363
},
{
"epoch": 0.5216061185468451,
"grad_norm": 5.296478526713344,
"learning_rate": 1e-06,
"loss": 0.6094,
"step": 1364
},
{
"epoch": 0.5219885277246654,
"grad_norm": 3.660781790612689,
"learning_rate": 1e-06,
"loss": 0.4412,
"step": 1365
},
{
"epoch": 0.5223709369024857,
"grad_norm": 1.9007224315694646,
"learning_rate": 1e-06,
"loss": 0.2527,
"step": 1366
},
{
"epoch": 0.5227533460803059,
"grad_norm": 2.956085658596427,
"learning_rate": 1e-06,
"loss": 0.231,
"step": 1367
},
{
"epoch": 0.5231357552581262,
"grad_norm": 2.4687345480133516,
"learning_rate": 1e-06,
"loss": 0.411,
"step": 1368
},
{
"epoch": 0.5235181644359465,
"grad_norm": 4.025379489992484,
"learning_rate": 1e-06,
"loss": 0.3887,
"step": 1369
},
{
"epoch": 0.5239005736137667,
"grad_norm": 3.6222658040311884,
"learning_rate": 1e-06,
"loss": 0.2028,
"step": 1370
},
{
"epoch": 0.524282982791587,
"grad_norm": 3.7365587305909913,
"learning_rate": 1e-06,
"loss": 0.2103,
"step": 1371
},
{
"epoch": 0.5246653919694073,
"grad_norm": 4.312172255293865,
"learning_rate": 1e-06,
"loss": 0.1642,
"step": 1372
},
{
"epoch": 0.5250478011472275,
"grad_norm": 3.3071052238397027,
"learning_rate": 1e-06,
"loss": 0.1017,
"step": 1373
},
{
"epoch": 0.5254302103250478,
"grad_norm": 4.405403651125026,
"learning_rate": 1e-06,
"loss": 0.1677,
"step": 1374
},
{
"epoch": 0.5258126195028681,
"grad_norm": 5.855008464126268,
"learning_rate": 1e-06,
"loss": 0.1313,
"step": 1375
},
{
"epoch": 0.5261950286806883,
"grad_norm": 10.89078227166302,
"learning_rate": 1e-06,
"loss": 0.5938,
"step": 1376
},
{
"epoch": 0.5265774378585086,
"grad_norm": 3.8335694502936306,
"learning_rate": 1e-06,
"loss": 0.3765,
"step": 1377
},
{
"epoch": 0.5269598470363289,
"grad_norm": 3.2972544953386977,
"learning_rate": 1e-06,
"loss": 0.6239,
"step": 1378
},
{
"epoch": 0.5273422562141491,
"grad_norm": 2.458371232166294,
"learning_rate": 1e-06,
"loss": 0.3414,
"step": 1379
},
{
"epoch": 0.5277246653919694,
"grad_norm": 2.390335564139875,
"learning_rate": 1e-06,
"loss": 0.5026,
"step": 1380
},
{
"epoch": 0.5281070745697897,
"grad_norm": 2.967322076414318,
"learning_rate": 1e-06,
"loss": 0.444,
"step": 1381
},
{
"epoch": 0.52848948374761,
"grad_norm": 2.8882291195297185,
"learning_rate": 1e-06,
"loss": 0.4309,
"step": 1382
},
{
"epoch": 0.5288718929254302,
"grad_norm": 3.138768800521931,
"learning_rate": 1e-06,
"loss": 0.2174,
"step": 1383
},
{
"epoch": 0.5292543021032505,
"grad_norm": 4.084695835252402,
"learning_rate": 1e-06,
"loss": 0.3231,
"step": 1384
},
{
"epoch": 0.5296367112810707,
"grad_norm": 6.997493840222035,
"learning_rate": 1e-06,
"loss": 0.3456,
"step": 1385
},
{
"epoch": 0.530019120458891,
"grad_norm": 4.038094085353025,
"learning_rate": 1e-06,
"loss": 0.17,
"step": 1386
},
{
"epoch": 0.5304015296367113,
"grad_norm": 5.436952234498637,
"learning_rate": 1e-06,
"loss": 0.1275,
"step": 1387
},
{
"epoch": 0.5307839388145316,
"grad_norm": 5.222542304291769,
"learning_rate": 1e-06,
"loss": 0.4724,
"step": 1388
},
{
"epoch": 0.5311663479923519,
"grad_norm": 5.941772850910443,
"learning_rate": 1e-06,
"loss": 0.4343,
"step": 1389
},
{
"epoch": 0.5315487571701721,
"grad_norm": 2.8879997910792734,
"learning_rate": 1e-06,
"loss": 0.3274,
"step": 1390
},
{
"epoch": 0.5319311663479923,
"grad_norm": 3.1758467138471143,
"learning_rate": 1e-06,
"loss": 0.5694,
"step": 1391
},
{
"epoch": 0.5323135755258126,
"grad_norm": 1.874729391279537,
"learning_rate": 1e-06,
"loss": 0.281,
"step": 1392
},
{
"epoch": 0.5326959847036329,
"grad_norm": 2.1564058993828286,
"learning_rate": 1e-06,
"loss": 0.3753,
"step": 1393
},
{
"epoch": 0.5330783938814532,
"grad_norm": 3.7039091028549707,
"learning_rate": 1e-06,
"loss": 0.4773,
"step": 1394
},
{
"epoch": 0.5334608030592735,
"grad_norm": 2.860529765435142,
"learning_rate": 1e-06,
"loss": 0.2912,
"step": 1395
},
{
"epoch": 0.5338432122370937,
"grad_norm": 2.727603047047569,
"learning_rate": 1e-06,
"loss": 0.2058,
"step": 1396
},
{
"epoch": 0.5342256214149139,
"grad_norm": 3.342870008236601,
"learning_rate": 1e-06,
"loss": 0.1445,
"step": 1397
},
{
"epoch": 0.5346080305927342,
"grad_norm": 6.050217288820573,
"learning_rate": 1e-06,
"loss": 0.1494,
"step": 1398
},
{
"epoch": 0.5349904397705545,
"grad_norm": 3.790319615679061,
"learning_rate": 1e-06,
"loss": 0.1409,
"step": 1399
},
{
"epoch": 0.5353728489483748,
"grad_norm": 5.62295698575686,
"learning_rate": 1e-06,
"loss": 0.2005,
"step": 1400
},
{
"epoch": 0.5357552581261951,
"grad_norm": 4.8012770543407575,
"learning_rate": 1e-06,
"loss": 0.335,
"step": 1401
},
{
"epoch": 0.5361376673040152,
"grad_norm": 3.8272160910788013,
"learning_rate": 1e-06,
"loss": 0.6847,
"step": 1402
},
{
"epoch": 0.5365200764818355,
"grad_norm": 1.7962285081271852,
"learning_rate": 1e-06,
"loss": 0.4374,
"step": 1403
},
{
"epoch": 0.5369024856596558,
"grad_norm": 2.3516397558556457,
"learning_rate": 1e-06,
"loss": 0.3764,
"step": 1404
},
{
"epoch": 0.5372848948374761,
"grad_norm": 4.221537255614015,
"learning_rate": 1e-06,
"loss": 0.603,
"step": 1405
},
{
"epoch": 0.5376673040152964,
"grad_norm": 3.0694035162481716,
"learning_rate": 1e-06,
"loss": 0.2605,
"step": 1406
},
{
"epoch": 0.5380497131931167,
"grad_norm": 3.531148318286807,
"learning_rate": 1e-06,
"loss": 0.2123,
"step": 1407
},
{
"epoch": 0.5384321223709368,
"grad_norm": 3.2479461635957287,
"learning_rate": 1e-06,
"loss": 0.3191,
"step": 1408
},
{
"epoch": 0.5388145315487571,
"grad_norm": 3.652223113816529,
"learning_rate": 1e-06,
"loss": 0.1757,
"step": 1409
},
{
"epoch": 0.5391969407265774,
"grad_norm": 3.214449185044371,
"learning_rate": 1e-06,
"loss": 0.2905,
"step": 1410
},
{
"epoch": 0.5395793499043977,
"grad_norm": 4.39707348340983,
"learning_rate": 1e-06,
"loss": 0.1222,
"step": 1411
},
{
"epoch": 0.539961759082218,
"grad_norm": 3.814813381045284,
"learning_rate": 1e-06,
"loss": 0.1384,
"step": 1412
},
{
"epoch": 0.5403441682600383,
"grad_norm": 5.64874882981047,
"learning_rate": 1e-06,
"loss": 0.2985,
"step": 1413
},
{
"epoch": 0.5407265774378585,
"grad_norm": 5.740154627933632,
"learning_rate": 1e-06,
"loss": 0.8276,
"step": 1414
},
{
"epoch": 0.5411089866156787,
"grad_norm": 3.4403655159623447,
"learning_rate": 1e-06,
"loss": 0.6347,
"step": 1415
},
{
"epoch": 0.541491395793499,
"grad_norm": 2.4875120114750535,
"learning_rate": 1e-06,
"loss": 0.3806,
"step": 1416
},
{
"epoch": 0.5418738049713193,
"grad_norm": 3.758326758966102,
"learning_rate": 1e-06,
"loss": 0.4696,
"step": 1417
},
{
"epoch": 0.5422562141491396,
"grad_norm": 5.878757979338721,
"learning_rate": 1e-06,
"loss": 0.4185,
"step": 1418
},
{
"epoch": 0.5426386233269599,
"grad_norm": 3.576439072998601,
"learning_rate": 1e-06,
"loss": 0.342,
"step": 1419
},
{
"epoch": 0.5430210325047801,
"grad_norm": 2.7456398123435597,
"learning_rate": 1e-06,
"loss": 0.2941,
"step": 1420
},
{
"epoch": 0.5434034416826004,
"grad_norm": 4.197590300160669,
"learning_rate": 1e-06,
"loss": 0.3272,
"step": 1421
},
{
"epoch": 0.5437858508604206,
"grad_norm": 3.2487043585805786,
"learning_rate": 1e-06,
"loss": 0.182,
"step": 1422
},
{
"epoch": 0.5441682600382409,
"grad_norm": 3.94322200571778,
"learning_rate": 1e-06,
"loss": 0.168,
"step": 1423
},
{
"epoch": 0.5445506692160612,
"grad_norm": 2.604046709476756,
"learning_rate": 1e-06,
"loss": 0.0688,
"step": 1424
},
{
"epoch": 0.5449330783938815,
"grad_norm": 3.660151038883697,
"learning_rate": 1e-06,
"loss": 0.1562,
"step": 1425
},
{
"epoch": 0.5453154875717017,
"grad_norm": 4.966302616849774,
"learning_rate": 1e-06,
"loss": 0.419,
"step": 1426
},
{
"epoch": 0.545697896749522,
"grad_norm": 4.336279283404338,
"learning_rate": 1e-06,
"loss": 0.5309,
"step": 1427
},
{
"epoch": 0.5460803059273422,
"grad_norm": 2.8745609445764884,
"learning_rate": 1e-06,
"loss": 0.5821,
"step": 1428
},
{
"epoch": 0.5464627151051625,
"grad_norm": 3.690098251233938,
"learning_rate": 1e-06,
"loss": 0.4779,
"step": 1429
},
{
"epoch": 0.5468451242829828,
"grad_norm": 3.2592457058974755,
"learning_rate": 1e-06,
"loss": 0.4654,
"step": 1430
},
{
"epoch": 0.5472275334608031,
"grad_norm": 2.274986761704379,
"learning_rate": 1e-06,
"loss": 0.2362,
"step": 1431
},
{
"epoch": 0.5476099426386233,
"grad_norm": 2.4617258898051313,
"learning_rate": 1e-06,
"loss": 0.3788,
"step": 1432
},
{
"epoch": 0.5479923518164436,
"grad_norm": 3.181641074614379,
"learning_rate": 1e-06,
"loss": 0.1856,
"step": 1433
},
{
"epoch": 0.5483747609942639,
"grad_norm": 3.587572709785552,
"learning_rate": 1e-06,
"loss": 0.2723,
"step": 1434
},
{
"epoch": 0.5487571701720841,
"grad_norm": 6.8930961347213024,
"learning_rate": 1e-06,
"loss": 0.1598,
"step": 1435
},
{
"epoch": 0.5491395793499044,
"grad_norm": 3.597531129067457,
"learning_rate": 1e-06,
"loss": 0.158,
"step": 1436
},
{
"epoch": 0.5495219885277247,
"grad_norm": 3.356841369471808,
"learning_rate": 1e-06,
"loss": 0.1753,
"step": 1437
},
{
"epoch": 0.5499043977055449,
"grad_norm": 5.611603445647346,
"learning_rate": 1e-06,
"loss": 0.3639,
"step": 1438
},
{
"epoch": 0.5502868068833652,
"grad_norm": 3.989556029436323,
"learning_rate": 1e-06,
"loss": 0.6409,
"step": 1439
},
{
"epoch": 0.5506692160611855,
"grad_norm": 3.3816313410443484,
"learning_rate": 1e-06,
"loss": 0.5594,
"step": 1440
},
{
"epoch": 0.5510516252390057,
"grad_norm": 2.4438288761297637,
"learning_rate": 1e-06,
"loss": 0.547,
"step": 1441
},
{
"epoch": 0.551434034416826,
"grad_norm": 2.419172193954764,
"learning_rate": 1e-06,
"loss": 0.4859,
"step": 1442
},
{
"epoch": 0.5518164435946463,
"grad_norm": 2.7785293473648034,
"learning_rate": 1e-06,
"loss": 0.3713,
"step": 1443
},
{
"epoch": 0.5521988527724665,
"grad_norm": 4.100156008845883,
"learning_rate": 1e-06,
"loss": 0.4,
"step": 1444
},
{
"epoch": 0.5525812619502868,
"grad_norm": 2.379213310180867,
"learning_rate": 1e-06,
"loss": 0.2778,
"step": 1445
},
{
"epoch": 0.5529636711281071,
"grad_norm": 3.3810549921292683,
"learning_rate": 1e-06,
"loss": 0.3165,
"step": 1446
},
{
"epoch": 0.5533460803059274,
"grad_norm": 3.1252262796494,
"learning_rate": 1e-06,
"loss": 0.2668,
"step": 1447
},
{
"epoch": 0.5537284894837476,
"grad_norm": 3.3292930594949097,
"learning_rate": 1e-06,
"loss": 0.1549,
"step": 1448
},
{
"epoch": 0.5541108986615679,
"grad_norm": 3.138692536490799,
"learning_rate": 1e-06,
"loss": 0.1964,
"step": 1449
},
{
"epoch": 0.5544933078393881,
"grad_norm": 3.1526886138033787,
"learning_rate": 1e-06,
"loss": 0.1309,
"step": 1450
},
{
"epoch": 0.5548757170172084,
"grad_norm": 5.321098101587102,
"learning_rate": 1e-06,
"loss": 0.4587,
"step": 1451
},
{
"epoch": 0.5552581261950287,
"grad_norm": 4.287469731438118,
"learning_rate": 1e-06,
"loss": 0.7825,
"step": 1452
},
{
"epoch": 0.555640535372849,
"grad_norm": 2.4409792595356885,
"learning_rate": 1e-06,
"loss": 0.274,
"step": 1453
},
{
"epoch": 0.5560229445506693,
"grad_norm": 3.1741668989869627,
"learning_rate": 1e-06,
"loss": 0.346,
"step": 1454
},
{
"epoch": 0.5564053537284895,
"grad_norm": 2.319172280410221,
"learning_rate": 1e-06,
"loss": 0.3755,
"step": 1455
},
{
"epoch": 0.5567877629063097,
"grad_norm": 3.0541576501592655,
"learning_rate": 1e-06,
"loss": 0.3353,
"step": 1456
},
{
"epoch": 0.55717017208413,
"grad_norm": 2.044743483667686,
"learning_rate": 1e-06,
"loss": 0.2576,
"step": 1457
},
{
"epoch": 0.5575525812619503,
"grad_norm": 3.212134367919686,
"learning_rate": 1e-06,
"loss": 0.2496,
"step": 1458
},
{
"epoch": 0.5579349904397706,
"grad_norm": 2.270169245590621,
"learning_rate": 1e-06,
"loss": 0.0757,
"step": 1459
},
{
"epoch": 0.5583173996175909,
"grad_norm": 3.051400760362656,
"learning_rate": 1e-06,
"loss": 0.1447,
"step": 1460
},
{
"epoch": 0.558699808795411,
"grad_norm": 4.732923832941204,
"learning_rate": 1e-06,
"loss": 0.1583,
"step": 1461
},
{
"epoch": 0.5590822179732313,
"grad_norm": 3.7772486930176847,
"learning_rate": 1e-06,
"loss": 0.1173,
"step": 1462
},
{
"epoch": 0.5594646271510516,
"grad_norm": 6.13316050714446,
"learning_rate": 1e-06,
"loss": 0.2376,
"step": 1463
},
{
"epoch": 0.5598470363288719,
"grad_norm": 3.706321270468106,
"learning_rate": 1e-06,
"loss": 0.842,
"step": 1464
},
{
"epoch": 0.5602294455066922,
"grad_norm": 3.6925066564211964,
"learning_rate": 1e-06,
"loss": 0.4967,
"step": 1465
},
{
"epoch": 0.5606118546845125,
"grad_norm": 3.049498382342764,
"learning_rate": 1e-06,
"loss": 0.3982,
"step": 1466
},
{
"epoch": 0.5609942638623326,
"grad_norm": 3.5211664226608517,
"learning_rate": 1e-06,
"loss": 0.5354,
"step": 1467
},
{
"epoch": 0.5613766730401529,
"grad_norm": 2.583751941933015,
"learning_rate": 1e-06,
"loss": 0.3475,
"step": 1468
},
{
"epoch": 0.5617590822179732,
"grad_norm": 3.5314702328223095,
"learning_rate": 1e-06,
"loss": 0.2992,
"step": 1469
},
{
"epoch": 0.5621414913957935,
"grad_norm": 2.7819103357165913,
"learning_rate": 1e-06,
"loss": 0.3423,
"step": 1470
},
{
"epoch": 0.5625239005736138,
"grad_norm": 2.6250383283450858,
"learning_rate": 1e-06,
"loss": 0.3856,
"step": 1471
},
{
"epoch": 0.5629063097514341,
"grad_norm": 3.511457898186983,
"learning_rate": 1e-06,
"loss": 0.3121,
"step": 1472
},
{
"epoch": 0.5632887189292543,
"grad_norm": 3.1110922513874555,
"learning_rate": 1e-06,
"loss": 0.1538,
"step": 1473
},
{
"epoch": 0.5636711281070745,
"grad_norm": 3.4822119597680246,
"learning_rate": 1e-06,
"loss": 0.1029,
"step": 1474
},
{
"epoch": 0.5640535372848948,
"grad_norm": 3.7911484472481773,
"learning_rate": 1e-06,
"loss": 0.1442,
"step": 1475
},
{
"epoch": 0.5644359464627151,
"grad_norm": 4.900222672054585,
"learning_rate": 1e-06,
"loss": 0.3698,
"step": 1476
},
{
"epoch": 0.5648183556405354,
"grad_norm": 3.617631856101012,
"learning_rate": 1e-06,
"loss": 0.7431,
"step": 1477
},
{
"epoch": 0.5652007648183557,
"grad_norm": 3.8977123495787365,
"learning_rate": 1e-06,
"loss": 0.3348,
"step": 1478
},
{
"epoch": 0.5655831739961759,
"grad_norm": 2.2151442766239327,
"learning_rate": 1e-06,
"loss": 0.4633,
"step": 1479
},
{
"epoch": 0.5659655831739961,
"grad_norm": 4.813007278459072,
"learning_rate": 1e-06,
"loss": 0.4659,
"step": 1480
},
{
"epoch": 0.5663479923518164,
"grad_norm": 3.7812468473563174,
"learning_rate": 1e-06,
"loss": 0.2921,
"step": 1481
},
{
"epoch": 0.5667304015296367,
"grad_norm": 2.1776800721929224,
"learning_rate": 1e-06,
"loss": 0.3112,
"step": 1482
},
{
"epoch": 0.567112810707457,
"grad_norm": 5.087591649385154,
"learning_rate": 1e-06,
"loss": 0.2409,
"step": 1483
},
{
"epoch": 0.5674952198852773,
"grad_norm": 3.8126374516958377,
"learning_rate": 1e-06,
"loss": 0.2257,
"step": 1484
},
{
"epoch": 0.5678776290630975,
"grad_norm": 4.73382625610079,
"learning_rate": 1e-06,
"loss": 0.2601,
"step": 1485
},
{
"epoch": 0.5682600382409178,
"grad_norm": 4.026684918983304,
"learning_rate": 1e-06,
"loss": 0.1744,
"step": 1486
},
{
"epoch": 0.568642447418738,
"grad_norm": 4.549533132558017,
"learning_rate": 1e-06,
"loss": 0.1276,
"step": 1487
},
{
"epoch": 0.5690248565965583,
"grad_norm": 5.288982441463886,
"learning_rate": 1e-06,
"loss": 0.3607,
"step": 1488
},
{
"epoch": 0.5694072657743786,
"grad_norm": 2.8286032930866507,
"learning_rate": 1e-06,
"loss": 0.3267,
"step": 1489
},
{
"epoch": 0.5697896749521989,
"grad_norm": 3.6822937241925757,
"learning_rate": 1e-06,
"loss": 0.4347,
"step": 1490
},
{
"epoch": 0.5701720841300191,
"grad_norm": 2.3247244148589967,
"learning_rate": 1e-06,
"loss": 0.554,
"step": 1491
},
{
"epoch": 0.5705544933078394,
"grad_norm": 2.3272486323767416,
"learning_rate": 1e-06,
"loss": 0.3814,
"step": 1492
},
{
"epoch": 0.5709369024856596,
"grad_norm": 3.7883676422709254,
"learning_rate": 1e-06,
"loss": 0.5862,
"step": 1493
},
{
"epoch": 0.5713193116634799,
"grad_norm": 3.1826315707138937,
"learning_rate": 1e-06,
"loss": 0.4448,
"step": 1494
},
{
"epoch": 0.5717017208413002,
"grad_norm": 3.4616438099733,
"learning_rate": 1e-06,
"loss": 0.3137,
"step": 1495
},
{
"epoch": 0.5720841300191205,
"grad_norm": 2.8849867811511705,
"learning_rate": 1e-06,
"loss": 0.2362,
"step": 1496
},
{
"epoch": 0.5724665391969407,
"grad_norm": 3.38657766926089,
"learning_rate": 1e-06,
"loss": 0.3042,
"step": 1497
},
{
"epoch": 0.572848948374761,
"grad_norm": 3.1036333142214962,
"learning_rate": 1e-06,
"loss": 0.3551,
"step": 1498
},
{
"epoch": 0.5732313575525813,
"grad_norm": 5.5639005462292435,
"learning_rate": 1e-06,
"loss": 0.2236,
"step": 1499
},
{
"epoch": 0.5736137667304015,
"grad_norm": 6.187496609157057,
"learning_rate": 1e-06,
"loss": 0.2325,
"step": 1500
},
{
"epoch": 0.5736137667304015,
"eval_runtime": 833.3537,
"eval_samples_per_second": 1.841,
"eval_steps_per_second": 0.461,
"step": 1500
},
{
"epoch": 0.5739961759082218,
"grad_norm": 7.648043223925274,
"learning_rate": 1e-06,
"loss": 0.4667,
"step": 1501
},
{
"epoch": 0.5743785850860421,
"grad_norm": 3.2906155577956144,
"learning_rate": 1e-06,
"loss": 0.4217,
"step": 1502
},
{
"epoch": 0.5747609942638623,
"grad_norm": 3.190112670056694,
"learning_rate": 1e-06,
"loss": 0.6407,
"step": 1503
},
{
"epoch": 0.5751434034416826,
"grad_norm": 2.420269046046126,
"learning_rate": 1e-06,
"loss": 0.2823,
"step": 1504
},
{
"epoch": 0.5755258126195029,
"grad_norm": 2.518074596333273,
"learning_rate": 1e-06,
"loss": 0.3212,
"step": 1505
},
{
"epoch": 0.5759082217973231,
"grad_norm": 3.2015445438773242,
"learning_rate": 1e-06,
"loss": 0.2174,
"step": 1506
},
{
"epoch": 0.5762906309751434,
"grad_norm": 3.1915377306147423,
"learning_rate": 1e-06,
"loss": 0.3889,
"step": 1507
},
{
"epoch": 0.5766730401529637,
"grad_norm": 2.5112891890343283,
"learning_rate": 1e-06,
"loss": 0.3093,
"step": 1508
},
{
"epoch": 0.5770554493307839,
"grad_norm": 2.5316295574683827,
"learning_rate": 1e-06,
"loss": 0.2095,
"step": 1509
},
{
"epoch": 0.5774378585086042,
"grad_norm": 5.745471331963637,
"learning_rate": 1e-06,
"loss": 0.332,
"step": 1510
},
{
"epoch": 0.5778202676864245,
"grad_norm": 4.534498852143188,
"learning_rate": 1e-06,
"loss": 0.1538,
"step": 1511
},
{
"epoch": 0.5782026768642448,
"grad_norm": 4.822514802189205,
"learning_rate": 1e-06,
"loss": 0.2601,
"step": 1512
},
{
"epoch": 0.578585086042065,
"grad_norm": 4.8454374265880755,
"learning_rate": 1e-06,
"loss": 0.3081,
"step": 1513
},
{
"epoch": 0.5789674952198853,
"grad_norm": 4.366637603041822,
"learning_rate": 1e-06,
"loss": 0.5815,
"step": 1514
},
{
"epoch": 0.5793499043977055,
"grad_norm": 4.221327156986129,
"learning_rate": 1e-06,
"loss": 0.5176,
"step": 1515
},
{
"epoch": 0.5797323135755258,
"grad_norm": 3.141811772915536,
"learning_rate": 1e-06,
"loss": 0.2951,
"step": 1516
},
{
"epoch": 0.5801147227533461,
"grad_norm": 2.4990048335144266,
"learning_rate": 1e-06,
"loss": 0.202,
"step": 1517
},
{
"epoch": 0.5804971319311664,
"grad_norm": 3.9050183605197164,
"learning_rate": 1e-06,
"loss": 0.4118,
"step": 1518
},
{
"epoch": 0.5808795411089867,
"grad_norm": 4.53393096537423,
"learning_rate": 1e-06,
"loss": 0.3508,
"step": 1519
},
{
"epoch": 0.5812619502868069,
"grad_norm": 2.961584176578633,
"learning_rate": 1e-06,
"loss": 0.2947,
"step": 1520
},
{
"epoch": 0.5816443594646271,
"grad_norm": 3.441416641895136,
"learning_rate": 1e-06,
"loss": 0.186,
"step": 1521
},
{
"epoch": 0.5820267686424474,
"grad_norm": 2.72530329355094,
"learning_rate": 1e-06,
"loss": 0.1401,
"step": 1522
},
{
"epoch": 0.5824091778202677,
"grad_norm": 2.818957227714633,
"learning_rate": 1e-06,
"loss": 0.2435,
"step": 1523
},
{
"epoch": 0.582791586998088,
"grad_norm": 4.097375792877866,
"learning_rate": 1e-06,
"loss": 0.1301,
"step": 1524
},
{
"epoch": 0.5831739961759083,
"grad_norm": 3.5816197548964563,
"learning_rate": 1e-06,
"loss": 0.1609,
"step": 1525
},
{
"epoch": 0.5835564053537284,
"grad_norm": 5.585264450820185,
"learning_rate": 1e-06,
"loss": 0.3654,
"step": 1526
},
{
"epoch": 0.5839388145315487,
"grad_norm": 4.122925670293326,
"learning_rate": 1e-06,
"loss": 0.5015,
"step": 1527
},
{
"epoch": 0.584321223709369,
"grad_norm": 2.4922913915331124,
"learning_rate": 1e-06,
"loss": 0.4,
"step": 1528
},
{
"epoch": 0.5847036328871893,
"grad_norm": 2.615770189637759,
"learning_rate": 1e-06,
"loss": 0.3547,
"step": 1529
},
{
"epoch": 0.5850860420650096,
"grad_norm": 2.193498398136122,
"learning_rate": 1e-06,
"loss": 0.2621,
"step": 1530
},
{
"epoch": 0.5854684512428299,
"grad_norm": 5.002734962142251,
"learning_rate": 1e-06,
"loss": 0.421,
"step": 1531
},
{
"epoch": 0.58585086042065,
"grad_norm": 3.724211841126314,
"learning_rate": 1e-06,
"loss": 0.3183,
"step": 1532
},
{
"epoch": 0.5862332695984703,
"grad_norm": 2.7640993899681088,
"learning_rate": 1e-06,
"loss": 0.4786,
"step": 1533
},
{
"epoch": 0.5866156787762906,
"grad_norm": 3.5616439326168727,
"learning_rate": 1e-06,
"loss": 0.241,
"step": 1534
},
{
"epoch": 0.5869980879541109,
"grad_norm": 2.4489084926408275,
"learning_rate": 1e-06,
"loss": 0.0949,
"step": 1535
},
{
"epoch": 0.5873804971319312,
"grad_norm": 3.6020111665102528,
"learning_rate": 1e-06,
"loss": 0.1548,
"step": 1536
},
{
"epoch": 0.5877629063097515,
"grad_norm": 3.941590379510361,
"learning_rate": 1e-06,
"loss": 0.1511,
"step": 1537
},
{
"epoch": 0.5881453154875717,
"grad_norm": 5.340405583213029,
"learning_rate": 1e-06,
"loss": 0.2523,
"step": 1538
},
{
"epoch": 0.5885277246653919,
"grad_norm": 4.8752568495162265,
"learning_rate": 1e-06,
"loss": 0.3552,
"step": 1539
},
{
"epoch": 0.5889101338432122,
"grad_norm": 3.6597323014379723,
"learning_rate": 1e-06,
"loss": 0.6053,
"step": 1540
},
{
"epoch": 0.5892925430210325,
"grad_norm": 3.8974402610404315,
"learning_rate": 1e-06,
"loss": 0.5765,
"step": 1541
},
{
"epoch": 0.5896749521988528,
"grad_norm": 2.729337749991348,
"learning_rate": 1e-06,
"loss": 0.2775,
"step": 1542
},
{
"epoch": 0.5900573613766731,
"grad_norm": 3.1248181099409362,
"learning_rate": 1e-06,
"loss": 0.2691,
"step": 1543
},
{
"epoch": 0.5904397705544933,
"grad_norm": 3.088190342872511,
"learning_rate": 1e-06,
"loss": 0.2857,
"step": 1544
},
{
"epoch": 0.5908221797323135,
"grad_norm": 2.6967299120653387,
"learning_rate": 1e-06,
"loss": 0.3731,
"step": 1545
},
{
"epoch": 0.5912045889101338,
"grad_norm": 2.647025392981962,
"learning_rate": 1e-06,
"loss": 0.2182,
"step": 1546
},
{
"epoch": 0.5915869980879541,
"grad_norm": 3.2712816226817005,
"learning_rate": 1e-06,
"loss": 0.2024,
"step": 1547
},
{
"epoch": 0.5919694072657744,
"grad_norm": 5.930421484752699,
"learning_rate": 1e-06,
"loss": 0.3072,
"step": 1548
},
{
"epoch": 0.5923518164435947,
"grad_norm": 3.686741508450827,
"learning_rate": 1e-06,
"loss": 0.1599,
"step": 1549
},
{
"epoch": 0.5927342256214149,
"grad_norm": 5.403013462307581,
"learning_rate": 1e-06,
"loss": 0.1649,
"step": 1550
},
{
"epoch": 0.5931166347992352,
"grad_norm": 5.0398586383552955,
"learning_rate": 1e-06,
"loss": 0.3195,
"step": 1551
},
{
"epoch": 0.5934990439770554,
"grad_norm": 4.363840748595389,
"learning_rate": 1e-06,
"loss": 0.6514,
"step": 1552
},
{
"epoch": 0.5938814531548757,
"grad_norm": 3.7887919221205557,
"learning_rate": 1e-06,
"loss": 0.5849,
"step": 1553
},
{
"epoch": 0.594263862332696,
"grad_norm": 3.225302831344963,
"learning_rate": 1e-06,
"loss": 0.2775,
"step": 1554
},
{
"epoch": 0.5946462715105163,
"grad_norm": 2.861391117587321,
"learning_rate": 1e-06,
"loss": 0.5969,
"step": 1555
},
{
"epoch": 0.5950286806883365,
"grad_norm": 2.9745260165601777,
"learning_rate": 1e-06,
"loss": 0.4259,
"step": 1556
},
{
"epoch": 0.5954110898661568,
"grad_norm": 3.123675408970642,
"learning_rate": 1e-06,
"loss": 0.4054,
"step": 1557
},
{
"epoch": 0.595793499043977,
"grad_norm": 2.1075426973879976,
"learning_rate": 1e-06,
"loss": 0.3542,
"step": 1558
},
{
"epoch": 0.5961759082217973,
"grad_norm": 2.500735556159103,
"learning_rate": 1e-06,
"loss": 0.2346,
"step": 1559
},
{
"epoch": 0.5965583173996176,
"grad_norm": 4.499995337589816,
"learning_rate": 1e-06,
"loss": 0.3501,
"step": 1560
},
{
"epoch": 0.5969407265774379,
"grad_norm": 3.784316655787347,
"learning_rate": 1e-06,
"loss": 0.215,
"step": 1561
},
{
"epoch": 0.5973231357552581,
"grad_norm": 2.7468269422057237,
"learning_rate": 1e-06,
"loss": 0.084,
"step": 1562
},
{
"epoch": 0.5977055449330784,
"grad_norm": 6.792289783312273,
"learning_rate": 1e-06,
"loss": 0.2943,
"step": 1563
},
{
"epoch": 0.5980879541108987,
"grad_norm": 3.762110926386478,
"learning_rate": 1e-06,
"loss": 0.4925,
"step": 1564
},
{
"epoch": 0.5984703632887189,
"grad_norm": 4.271277626312197,
"learning_rate": 1e-06,
"loss": 0.6514,
"step": 1565
},
{
"epoch": 0.5988527724665392,
"grad_norm": 4.796524681197209,
"learning_rate": 1e-06,
"loss": 0.17,
"step": 1566
},
{
"epoch": 0.5992351816443595,
"grad_norm": 2.6530873937379136,
"learning_rate": 1e-06,
"loss": 0.4564,
"step": 1567
},
{
"epoch": 0.5996175908221797,
"grad_norm": 3.284529146448116,
"learning_rate": 1e-06,
"loss": 0.4594,
"step": 1568
},
{
"epoch": 0.6,
"grad_norm": 4.297701669413807,
"learning_rate": 1e-06,
"loss": 0.5233,
"step": 1569
},
{
"epoch": 0.6003824091778203,
"grad_norm": 2.7848361608503187,
"learning_rate": 1e-06,
"loss": 0.2024,
"step": 1570
},
{
"epoch": 0.6007648183556406,
"grad_norm": 3.1810191971967994,
"learning_rate": 1e-06,
"loss": 0.2397,
"step": 1571
},
{
"epoch": 0.6011472275334608,
"grad_norm": 3.462055516547677,
"learning_rate": 1e-06,
"loss": 0.2973,
"step": 1572
},
{
"epoch": 0.6015296367112811,
"grad_norm": 3.9998092605890285,
"learning_rate": 1e-06,
"loss": 0.2522,
"step": 1573
},
{
"epoch": 0.6019120458891013,
"grad_norm": 3.0841793411792917,
"learning_rate": 1e-06,
"loss": 0.0995,
"step": 1574
},
{
"epoch": 0.6022944550669216,
"grad_norm": 4.263103256718209,
"learning_rate": 1e-06,
"loss": 0.122,
"step": 1575
},
{
"epoch": 0.6026768642447419,
"grad_norm": 6.395907357084697,
"learning_rate": 1e-06,
"loss": 0.4544,
"step": 1576
},
{
"epoch": 0.6030592734225622,
"grad_norm": 4.479918920941317,
"learning_rate": 1e-06,
"loss": 0.592,
"step": 1577
},
{
"epoch": 0.6034416826003824,
"grad_norm": 3.0270337542897297,
"learning_rate": 1e-06,
"loss": 0.3419,
"step": 1578
},
{
"epoch": 0.6038240917782027,
"grad_norm": 1.8534895128314122,
"learning_rate": 1e-06,
"loss": 0.3151,
"step": 1579
},
{
"epoch": 0.6042065009560229,
"grad_norm": 2.6555878655433514,
"learning_rate": 1e-06,
"loss": 0.6539,
"step": 1580
},
{
"epoch": 0.6045889101338432,
"grad_norm": 2.056482727423432,
"learning_rate": 1e-06,
"loss": 0.3516,
"step": 1581
},
{
"epoch": 0.6049713193116635,
"grad_norm": 3.275965913625526,
"learning_rate": 1e-06,
"loss": 0.1933,
"step": 1582
},
{
"epoch": 0.6053537284894838,
"grad_norm": 3.0842731864844897,
"learning_rate": 1e-06,
"loss": 0.3483,
"step": 1583
},
{
"epoch": 0.605736137667304,
"grad_norm": 2.7037476280085597,
"learning_rate": 1e-06,
"loss": 0.2797,
"step": 1584
},
{
"epoch": 0.6061185468451242,
"grad_norm": 3.385057792226577,
"learning_rate": 1e-06,
"loss": 0.1719,
"step": 1585
},
{
"epoch": 0.6065009560229445,
"grad_norm": 3.4903293248748812,
"learning_rate": 1e-06,
"loss": 0.1536,
"step": 1586
},
{
"epoch": 0.6068833652007648,
"grad_norm": 4.144413895163444,
"learning_rate": 1e-06,
"loss": 0.1619,
"step": 1587
},
{
"epoch": 0.6072657743785851,
"grad_norm": 4.818634680964997,
"learning_rate": 1e-06,
"loss": 0.2627,
"step": 1588
},
{
"epoch": 0.6076481835564054,
"grad_norm": 4.040560829968868,
"learning_rate": 1e-06,
"loss": 0.4934,
"step": 1589
},
{
"epoch": 0.6080305927342257,
"grad_norm": 4.085364219462454,
"learning_rate": 1e-06,
"loss": 0.5564,
"step": 1590
},
{
"epoch": 0.6084130019120458,
"grad_norm": 2.268925965265979,
"learning_rate": 1e-06,
"loss": 0.1791,
"step": 1591
},
{
"epoch": 0.6087954110898661,
"grad_norm": 3.289065254554592,
"learning_rate": 1e-06,
"loss": 0.2348,
"step": 1592
},
{
"epoch": 0.6091778202676864,
"grad_norm": 1.9265604357928798,
"learning_rate": 1e-06,
"loss": 0.3429,
"step": 1593
},
{
"epoch": 0.6095602294455067,
"grad_norm": 6.16191158309742,
"learning_rate": 1e-06,
"loss": 0.6237,
"step": 1594
},
{
"epoch": 0.609942638623327,
"grad_norm": 2.846177793984969,
"learning_rate": 1e-06,
"loss": 0.3524,
"step": 1595
},
{
"epoch": 0.6103250478011473,
"grad_norm": 3.2424341314797536,
"learning_rate": 1e-06,
"loss": 0.1905,
"step": 1596
},
{
"epoch": 0.6107074569789674,
"grad_norm": 2.762463071412313,
"learning_rate": 1e-06,
"loss": 0.309,
"step": 1597
},
{
"epoch": 0.6110898661567877,
"grad_norm": 5.149475085719637,
"learning_rate": 1e-06,
"loss": 0.1613,
"step": 1598
},
{
"epoch": 0.611472275334608,
"grad_norm": 2.5937691193760686,
"learning_rate": 1e-06,
"loss": 0.0886,
"step": 1599
},
{
"epoch": 0.6118546845124283,
"grad_norm": 4.107008335919648,
"learning_rate": 1e-06,
"loss": 0.102,
"step": 1600
},
{
"epoch": 0.6122370936902486,
"grad_norm": 5.038598232912599,
"learning_rate": 1e-06,
"loss": 0.4016,
"step": 1601
},
{
"epoch": 0.6126195028680689,
"grad_norm": 3.6673026978221492,
"learning_rate": 1e-06,
"loss": 0.5224,
"step": 1602
},
{
"epoch": 0.613001912045889,
"grad_norm": 1.9355971313480531,
"learning_rate": 1e-06,
"loss": 0.4279,
"step": 1603
},
{
"epoch": 0.6133843212237093,
"grad_norm": 3.028508942851441,
"learning_rate": 1e-06,
"loss": 0.3999,
"step": 1604
},
{
"epoch": 0.6137667304015296,
"grad_norm": 3.490329461491545,
"learning_rate": 1e-06,
"loss": 0.4828,
"step": 1605
},
{
"epoch": 0.6141491395793499,
"grad_norm": 3.799259143687563,
"learning_rate": 1e-06,
"loss": 0.4321,
"step": 1606
},
{
"epoch": 0.6145315487571702,
"grad_norm": 3.293860915341487,
"learning_rate": 1e-06,
"loss": 0.3982,
"step": 1607
},
{
"epoch": 0.6149139579349905,
"grad_norm": 3.2708085822527555,
"learning_rate": 1e-06,
"loss": 0.2293,
"step": 1608
},
{
"epoch": 0.6152963671128107,
"grad_norm": 3.449351600862056,
"learning_rate": 1e-06,
"loss": 0.358,
"step": 1609
},
{
"epoch": 0.615678776290631,
"grad_norm": 4.529874816579575,
"learning_rate": 1e-06,
"loss": 0.1249,
"step": 1610
},
{
"epoch": 0.6160611854684512,
"grad_norm": 3.6972732342351042,
"learning_rate": 1e-06,
"loss": 0.1118,
"step": 1611
},
{
"epoch": 0.6164435946462715,
"grad_norm": 5.522058468562662,
"learning_rate": 1e-06,
"loss": 0.1054,
"step": 1612
},
{
"epoch": 0.6168260038240918,
"grad_norm": 7.399685167368398,
"learning_rate": 1e-06,
"loss": 0.4509,
"step": 1613
},
{
"epoch": 0.6172084130019121,
"grad_norm": 4.929145021409884,
"learning_rate": 1e-06,
"loss": 0.5219,
"step": 1614
},
{
"epoch": 0.6175908221797323,
"grad_norm": 3.370637546758573,
"learning_rate": 1e-06,
"loss": 0.5646,
"step": 1615
},
{
"epoch": 0.6179732313575526,
"grad_norm": 2.3358637054932956,
"learning_rate": 1e-06,
"loss": 0.3913,
"step": 1616
},
{
"epoch": 0.6183556405353728,
"grad_norm": 3.445634407350294,
"learning_rate": 1e-06,
"loss": 0.4011,
"step": 1617
},
{
"epoch": 0.6187380497131931,
"grad_norm": 3.4811639744318743,
"learning_rate": 1e-06,
"loss": 0.2578,
"step": 1618
},
{
"epoch": 0.6191204588910134,
"grad_norm": 4.069633441886181,
"learning_rate": 1e-06,
"loss": 0.299,
"step": 1619
},
{
"epoch": 0.6195028680688337,
"grad_norm": 2.0427357994877333,
"learning_rate": 1e-06,
"loss": 0.124,
"step": 1620
},
{
"epoch": 0.6198852772466539,
"grad_norm": 3.318965432166625,
"learning_rate": 1e-06,
"loss": 0.3652,
"step": 1621
},
{
"epoch": 0.6202676864244742,
"grad_norm": 3.1058845289668113,
"learning_rate": 1e-06,
"loss": 0.1816,
"step": 1622
},
{
"epoch": 0.6206500956022944,
"grad_norm": 3.494146083701717,
"learning_rate": 1e-06,
"loss": 0.2191,
"step": 1623
},
{
"epoch": 0.6210325047801147,
"grad_norm": 2.911256968586936,
"learning_rate": 1e-06,
"loss": 0.0933,
"step": 1624
},
{
"epoch": 0.621414913957935,
"grad_norm": 4.57964187854441,
"learning_rate": 1e-06,
"loss": 0.2255,
"step": 1625
},
{
"epoch": 0.6217973231357553,
"grad_norm": 7.234922581952464,
"learning_rate": 1e-06,
"loss": 0.4494,
"step": 1626
},
{
"epoch": 0.6221797323135755,
"grad_norm": 2.7171413880571014,
"learning_rate": 1e-06,
"loss": 0.5141,
"step": 1627
},
{
"epoch": 0.6225621414913958,
"grad_norm": 3.4219758876751425,
"learning_rate": 1e-06,
"loss": 0.3405,
"step": 1628
},
{
"epoch": 0.6229445506692161,
"grad_norm": 2.5694710417966085,
"learning_rate": 1e-06,
"loss": 0.395,
"step": 1629
},
{
"epoch": 0.6233269598470363,
"grad_norm": 3.0536773650517013,
"learning_rate": 1e-06,
"loss": 0.2961,
"step": 1630
},
{
"epoch": 0.6237093690248566,
"grad_norm": 3.8657433481687753,
"learning_rate": 1e-06,
"loss": 0.3545,
"step": 1631
},
{
"epoch": 0.6240917782026769,
"grad_norm": 3.1711600865062755,
"learning_rate": 1e-06,
"loss": 0.3322,
"step": 1632
},
{
"epoch": 0.6244741873804971,
"grad_norm": 2.5303667199251354,
"learning_rate": 1e-06,
"loss": 0.3185,
"step": 1633
},
{
"epoch": 0.6248565965583174,
"grad_norm": 3.089084536792682,
"learning_rate": 1e-06,
"loss": 0.2415,
"step": 1634
},
{
"epoch": 0.6252390057361377,
"grad_norm": 5.177332093718016,
"learning_rate": 1e-06,
"loss": 0.196,
"step": 1635
},
{
"epoch": 0.625621414913958,
"grad_norm": 2.9635470121544407,
"learning_rate": 1e-06,
"loss": 0.1352,
"step": 1636
},
{
"epoch": 0.6260038240917782,
"grad_norm": 3.3452187993175597,
"learning_rate": 1e-06,
"loss": 0.1187,
"step": 1637
},
{
"epoch": 0.6263862332695985,
"grad_norm": 6.97572531808549,
"learning_rate": 1e-06,
"loss": 0.3908,
"step": 1638
},
{
"epoch": 0.6267686424474187,
"grad_norm": 5.440352820567418,
"learning_rate": 1e-06,
"loss": 0.6388,
"step": 1639
},
{
"epoch": 0.627151051625239,
"grad_norm": 2.9229332950379163,
"learning_rate": 1e-06,
"loss": 0.4336,
"step": 1640
},
{
"epoch": 0.6275334608030593,
"grad_norm": 2.539355451849907,
"learning_rate": 1e-06,
"loss": 0.3208,
"step": 1641
},
{
"epoch": 0.6279158699808796,
"grad_norm": 2.289203105683247,
"learning_rate": 1e-06,
"loss": 0.4489,
"step": 1642
},
{
"epoch": 0.6282982791586998,
"grad_norm": 3.1309029331615954,
"learning_rate": 1e-06,
"loss": 0.3362,
"step": 1643
},
{
"epoch": 0.62868068833652,
"grad_norm": 3.3206080406153653,
"learning_rate": 1e-06,
"loss": 0.1911,
"step": 1644
},
{
"epoch": 0.6290630975143403,
"grad_norm": 3.1214176243879708,
"learning_rate": 1e-06,
"loss": 0.3695,
"step": 1645
},
{
"epoch": 0.6294455066921606,
"grad_norm": 3.8474336666265994,
"learning_rate": 1e-06,
"loss": 0.1746,
"step": 1646
},
{
"epoch": 0.6298279158699809,
"grad_norm": 3.2207731536599136,
"learning_rate": 1e-06,
"loss": 0.1577,
"step": 1647
},
{
"epoch": 0.6302103250478012,
"grad_norm": 2.067440358308397,
"learning_rate": 1e-06,
"loss": 0.1543,
"step": 1648
},
{
"epoch": 0.6305927342256215,
"grad_norm": 2.2835514530593635,
"learning_rate": 1e-06,
"loss": 0.0666,
"step": 1649
},
{
"epoch": 0.6309751434034416,
"grad_norm": 4.402501669541691,
"learning_rate": 1e-06,
"loss": 0.1261,
"step": 1650
},
{
"epoch": 0.6313575525812619,
"grad_norm": 5.3180156124555795,
"learning_rate": 1e-06,
"loss": 0.34,
"step": 1651
},
{
"epoch": 0.6317399617590822,
"grad_norm": 4.015899292598844,
"learning_rate": 1e-06,
"loss": 0.3995,
"step": 1652
},
{
"epoch": 0.6321223709369025,
"grad_norm": 2.813014089435832,
"learning_rate": 1e-06,
"loss": 0.5354,
"step": 1653
},
{
"epoch": 0.6325047801147228,
"grad_norm": 1.4126292599974228,
"learning_rate": 1e-06,
"loss": 0.5314,
"step": 1654
},
{
"epoch": 0.6328871892925431,
"grad_norm": 2.7594537363019023,
"learning_rate": 1e-06,
"loss": 0.5301,
"step": 1655
},
{
"epoch": 0.6332695984703632,
"grad_norm": 2.8815974954834624,
"learning_rate": 1e-06,
"loss": 0.3094,
"step": 1656
},
{
"epoch": 0.6336520076481835,
"grad_norm": 2.462411106267252,
"learning_rate": 1e-06,
"loss": 0.2472,
"step": 1657
},
{
"epoch": 0.6340344168260038,
"grad_norm": 3.105835399935763,
"learning_rate": 1e-06,
"loss": 0.3324,
"step": 1658
},
{
"epoch": 0.6344168260038241,
"grad_norm": 5.450155679421611,
"learning_rate": 1e-06,
"loss": 0.1916,
"step": 1659
},
{
"epoch": 0.6347992351816444,
"grad_norm": 5.199921225904796,
"learning_rate": 1e-06,
"loss": 0.1978,
"step": 1660
},
{
"epoch": 0.6351816443594647,
"grad_norm": 3.691235476787247,
"learning_rate": 1e-06,
"loss": 0.1718,
"step": 1661
},
{
"epoch": 0.6355640535372848,
"grad_norm": 5.067963089637659,
"learning_rate": 1e-06,
"loss": 0.1273,
"step": 1662
},
{
"epoch": 0.6359464627151051,
"grad_norm": 5.3491196451746665,
"learning_rate": 1e-06,
"loss": 0.3786,
"step": 1663
},
{
"epoch": 0.6363288718929254,
"grad_norm": 4.173036106563145,
"learning_rate": 1e-06,
"loss": 0.4033,
"step": 1664
},
{
"epoch": 0.6367112810707457,
"grad_norm": 3.6322536900100064,
"learning_rate": 1e-06,
"loss": 0.5193,
"step": 1665
},
{
"epoch": 0.637093690248566,
"grad_norm": 2.4756148297130585,
"learning_rate": 1e-06,
"loss": 0.4113,
"step": 1666
},
{
"epoch": 0.6374760994263863,
"grad_norm": 3.255770914927566,
"learning_rate": 1e-06,
"loss": 0.3114,
"step": 1667
},
{
"epoch": 0.6378585086042065,
"grad_norm": 3.8464504773120454,
"learning_rate": 1e-06,
"loss": 0.4712,
"step": 1668
},
{
"epoch": 0.6382409177820267,
"grad_norm": 2.955989195467105,
"learning_rate": 1e-06,
"loss": 0.1994,
"step": 1669
},
{
"epoch": 0.638623326959847,
"grad_norm": 2.600741970124298,
"learning_rate": 1e-06,
"loss": 0.1038,
"step": 1670
},
{
"epoch": 0.6390057361376673,
"grad_norm": 3.1099354607855556,
"learning_rate": 1e-06,
"loss": 0.2166,
"step": 1671
},
{
"epoch": 0.6393881453154876,
"grad_norm": 3.5664288456379634,
"learning_rate": 1e-06,
"loss": 0.3254,
"step": 1672
},
{
"epoch": 0.6397705544933079,
"grad_norm": 3.727625281456364,
"learning_rate": 1e-06,
"loss": 0.2059,
"step": 1673
},
{
"epoch": 0.6401529636711281,
"grad_norm": 3.330111631521315,
"learning_rate": 1e-06,
"loss": 0.1298,
"step": 1674
},
{
"epoch": 0.6405353728489483,
"grad_norm": 5.150736872723286,
"learning_rate": 1e-06,
"loss": 0.1979,
"step": 1675
},
{
"epoch": 0.6409177820267686,
"grad_norm": 4.3827972683837375,
"learning_rate": 1e-06,
"loss": 0.4554,
"step": 1676
},
{
"epoch": 0.6413001912045889,
"grad_norm": 5.885968945902527,
"learning_rate": 1e-06,
"loss": 0.4346,
"step": 1677
},
{
"epoch": 0.6416826003824092,
"grad_norm": 2.1649212286805266,
"learning_rate": 1e-06,
"loss": 0.488,
"step": 1678
},
{
"epoch": 0.6420650095602295,
"grad_norm": 2.8184786511597806,
"learning_rate": 1e-06,
"loss": 0.3281,
"step": 1679
},
{
"epoch": 0.6424474187380497,
"grad_norm": 2.618282624192141,
"learning_rate": 1e-06,
"loss": 0.3105,
"step": 1680
},
{
"epoch": 0.64282982791587,
"grad_norm": 2.867418825552063,
"learning_rate": 1e-06,
"loss": 0.2497,
"step": 1681
},
{
"epoch": 0.6432122370936902,
"grad_norm": 2.421551885892965,
"learning_rate": 1e-06,
"loss": 0.3143,
"step": 1682
},
{
"epoch": 0.6435946462715105,
"grad_norm": 2.664223942940917,
"learning_rate": 1e-06,
"loss": 0.1734,
"step": 1683
},
{
"epoch": 0.6439770554493308,
"grad_norm": 2.357248118143426,
"learning_rate": 1e-06,
"loss": 0.1671,
"step": 1684
},
{
"epoch": 0.6443594646271511,
"grad_norm": 6.723963502329862,
"learning_rate": 1e-06,
"loss": 0.2437,
"step": 1685
},
{
"epoch": 0.6447418738049713,
"grad_norm": 5.0651138527850295,
"learning_rate": 1e-06,
"loss": 0.2618,
"step": 1686
},
{
"epoch": 0.6451242829827916,
"grad_norm": 4.32736849456671,
"learning_rate": 1e-06,
"loss": 0.1156,
"step": 1687
},
{
"epoch": 0.6455066921606119,
"grad_norm": 5.403851986058783,
"learning_rate": 1e-06,
"loss": 0.2925,
"step": 1688
},
{
"epoch": 0.6458891013384321,
"grad_norm": 5.4870437258588804,
"learning_rate": 1e-06,
"loss": 0.5704,
"step": 1689
},
{
"epoch": 0.6462715105162524,
"grad_norm": 2.9527269529089644,
"learning_rate": 1e-06,
"loss": 0.4319,
"step": 1690
},
{
"epoch": 0.6466539196940727,
"grad_norm": 2.7521387799722774,
"learning_rate": 1e-06,
"loss": 0.5618,
"step": 1691
},
{
"epoch": 0.6470363288718929,
"grad_norm": 3.840974131049391,
"learning_rate": 1e-06,
"loss": 0.3291,
"step": 1692
},
{
"epoch": 0.6474187380497132,
"grad_norm": 3.426352719670689,
"learning_rate": 1e-06,
"loss": 0.5471,
"step": 1693
},
{
"epoch": 0.6478011472275335,
"grad_norm": 4.976341733721796,
"learning_rate": 1e-06,
"loss": 0.4847,
"step": 1694
},
{
"epoch": 0.6481835564053537,
"grad_norm": 3.552568478981308,
"learning_rate": 1e-06,
"loss": 0.43,
"step": 1695
},
{
"epoch": 0.648565965583174,
"grad_norm": 2.934198512740815,
"learning_rate": 1e-06,
"loss": 0.3435,
"step": 1696
},
{
"epoch": 0.6489483747609943,
"grad_norm": 2.8276520222911943,
"learning_rate": 1e-06,
"loss": 0.2352,
"step": 1697
},
{
"epoch": 0.6493307839388145,
"grad_norm": 3.9218239343973753,
"learning_rate": 1e-06,
"loss": 0.2914,
"step": 1698
},
{
"epoch": 0.6497131931166348,
"grad_norm": 3.3779882694490158,
"learning_rate": 1e-06,
"loss": 0.1081,
"step": 1699
},
{
"epoch": 0.6500956022944551,
"grad_norm": 4.367757852123347,
"learning_rate": 1e-06,
"loss": 0.1843,
"step": 1700
},
{
"epoch": 0.6504780114722754,
"grad_norm": 5.006101890401848,
"learning_rate": 1e-06,
"loss": 0.7616,
"step": 1701
},
{
"epoch": 0.6508604206500956,
"grad_norm": 3.2721356919280042,
"learning_rate": 1e-06,
"loss": 0.2695,
"step": 1702
},
{
"epoch": 0.6512428298279158,
"grad_norm": 1.3918892760416124,
"learning_rate": 1e-06,
"loss": 0.5447,
"step": 1703
},
{
"epoch": 0.6516252390057361,
"grad_norm": 3.9619045544238265,
"learning_rate": 1e-06,
"loss": 0.7589,
"step": 1704
},
{
"epoch": 0.6520076481835564,
"grad_norm": 2.9728468083222777,
"learning_rate": 1e-06,
"loss": 0.2531,
"step": 1705
},
{
"epoch": 0.6523900573613767,
"grad_norm": 2.996392624332678,
"learning_rate": 1e-06,
"loss": 0.3013,
"step": 1706
},
{
"epoch": 0.652772466539197,
"grad_norm": 3.7618078615402455,
"learning_rate": 1e-06,
"loss": 0.2693,
"step": 1707
},
{
"epoch": 0.6531548757170172,
"grad_norm": 2.285208729198416,
"learning_rate": 1e-06,
"loss": 0.1972,
"step": 1708
},
{
"epoch": 0.6535372848948374,
"grad_norm": 3.874416184284152,
"learning_rate": 1e-06,
"loss": 0.1978,
"step": 1709
},
{
"epoch": 0.6539196940726577,
"grad_norm": 3.1600430801931734,
"learning_rate": 1e-06,
"loss": 0.1301,
"step": 1710
},
{
"epoch": 0.654302103250478,
"grad_norm": 5.052248524065872,
"learning_rate": 1e-06,
"loss": 0.3586,
"step": 1711
},
{
"epoch": 0.6546845124282983,
"grad_norm": 2.662989425449097,
"learning_rate": 1e-06,
"loss": 0.09,
"step": 1712
},
{
"epoch": 0.6550669216061186,
"grad_norm": 6.50430067544631,
"learning_rate": 1e-06,
"loss": 0.3388,
"step": 1713
},
{
"epoch": 0.6554493307839389,
"grad_norm": 3.6709607873296353,
"learning_rate": 1e-06,
"loss": 0.4649,
"step": 1714
},
{
"epoch": 0.655831739961759,
"grad_norm": 3.3219462605079326,
"learning_rate": 1e-06,
"loss": 0.4879,
"step": 1715
},
{
"epoch": 0.6562141491395793,
"grad_norm": 3.2695636269118244,
"learning_rate": 1e-06,
"loss": 0.5663,
"step": 1716
},
{
"epoch": 0.6565965583173996,
"grad_norm": 2.92610914279973,
"learning_rate": 1e-06,
"loss": 0.5111,
"step": 1717
},
{
"epoch": 0.6569789674952199,
"grad_norm": 4.1060225007137,
"learning_rate": 1e-06,
"loss": 0.5329,
"step": 1718
},
{
"epoch": 0.6573613766730402,
"grad_norm": 3.1458467563233303,
"learning_rate": 1e-06,
"loss": 0.3792,
"step": 1719
},
{
"epoch": 0.6577437858508605,
"grad_norm": 2.8146208502026533,
"learning_rate": 1e-06,
"loss": 0.2088,
"step": 1720
},
{
"epoch": 0.6581261950286806,
"grad_norm": 2.5008707437468076,
"learning_rate": 1e-06,
"loss": 0.375,
"step": 1721
},
{
"epoch": 0.6585086042065009,
"grad_norm": 4.107443700161939,
"learning_rate": 1e-06,
"loss": 0.17,
"step": 1722
},
{
"epoch": 0.6588910133843212,
"grad_norm": 3.5735843283007958,
"learning_rate": 1e-06,
"loss": 0.2203,
"step": 1723
},
{
"epoch": 0.6592734225621415,
"grad_norm": 3.2595638992124876,
"learning_rate": 1e-06,
"loss": 0.1531,
"step": 1724
},
{
"epoch": 0.6596558317399618,
"grad_norm": 3.4207497057215885,
"learning_rate": 1e-06,
"loss": 0.1183,
"step": 1725
},
{
"epoch": 0.6600382409177821,
"grad_norm": 4.449081994852851,
"learning_rate": 1e-06,
"loss": 0.4619,
"step": 1726
},
{
"epoch": 0.6604206500956022,
"grad_norm": 3.645980105397598,
"learning_rate": 1e-06,
"loss": 0.5698,
"step": 1727
},
{
"epoch": 0.6608030592734225,
"grad_norm": 3.17886019354846,
"learning_rate": 1e-06,
"loss": 0.576,
"step": 1728
},
{
"epoch": 0.6611854684512428,
"grad_norm": 2.309275982996253,
"learning_rate": 1e-06,
"loss": 0.26,
"step": 1729
},
{
"epoch": 0.6615678776290631,
"grad_norm": 3.1960452855402597,
"learning_rate": 1e-06,
"loss": 0.24,
"step": 1730
},
{
"epoch": 0.6619502868068834,
"grad_norm": 2.987791013709219,
"learning_rate": 1e-06,
"loss": 0.3228,
"step": 1731
},
{
"epoch": 0.6623326959847037,
"grad_norm": 4.365707774596113,
"learning_rate": 1e-06,
"loss": 0.3277,
"step": 1732
},
{
"epoch": 0.6627151051625239,
"grad_norm": 2.2909913773160038,
"learning_rate": 1e-06,
"loss": 0.1836,
"step": 1733
},
{
"epoch": 0.6630975143403441,
"grad_norm": 2.5442436058021807,
"learning_rate": 1e-06,
"loss": 0.2696,
"step": 1734
},
{
"epoch": 0.6634799235181644,
"grad_norm": 3.717153511036284,
"learning_rate": 1e-06,
"loss": 0.1953,
"step": 1735
},
{
"epoch": 0.6638623326959847,
"grad_norm": 3.883942898127248,
"learning_rate": 1e-06,
"loss": 0.1044,
"step": 1736
},
{
"epoch": 0.664244741873805,
"grad_norm": 4.682744984081185,
"learning_rate": 1e-06,
"loss": 0.185,
"step": 1737
},
{
"epoch": 0.6646271510516253,
"grad_norm": 6.51903592851396,
"learning_rate": 1e-06,
"loss": 0.2527,
"step": 1738
},
{
"epoch": 0.6650095602294455,
"grad_norm": 4.508646816747134,
"learning_rate": 1e-06,
"loss": 0.4237,
"step": 1739
},
{
"epoch": 0.6653919694072657,
"grad_norm": 4.032820994306484,
"learning_rate": 1e-06,
"loss": 0.3399,
"step": 1740
},
{
"epoch": 0.665774378585086,
"grad_norm": 2.9156831627182824,
"learning_rate": 1e-06,
"loss": 0.5586,
"step": 1741
},
{
"epoch": 0.6661567877629063,
"grad_norm": 2.5427943578595387,
"learning_rate": 1e-06,
"loss": 0.2631,
"step": 1742
},
{
"epoch": 0.6665391969407266,
"grad_norm": 3.71340905051251,
"learning_rate": 1e-06,
"loss": 0.6265,
"step": 1743
},
{
"epoch": 0.6669216061185469,
"grad_norm": 3.0503382635928866,
"learning_rate": 1e-06,
"loss": 0.322,
"step": 1744
},
{
"epoch": 0.6673040152963671,
"grad_norm": 3.1499401510699,
"learning_rate": 1e-06,
"loss": 0.3289,
"step": 1745
},
{
"epoch": 0.6676864244741874,
"grad_norm": 3.863479470316066,
"learning_rate": 1e-06,
"loss": 0.2829,
"step": 1746
},
{
"epoch": 0.6680688336520076,
"grad_norm": 3.4202907849106525,
"learning_rate": 1e-06,
"loss": 0.3704,
"step": 1747
},
{
"epoch": 0.6684512428298279,
"grad_norm": 2.5903809509547693,
"learning_rate": 1e-06,
"loss": 0.1806,
"step": 1748
},
{
"epoch": 0.6688336520076482,
"grad_norm": 3.983156264664412,
"learning_rate": 1e-06,
"loss": 0.0728,
"step": 1749
},
{
"epoch": 0.6692160611854685,
"grad_norm": 3.8396316423169607,
"learning_rate": 1e-06,
"loss": 0.1351,
"step": 1750
},
{
"epoch": 0.6695984703632887,
"grad_norm": 6.639303636918794,
"learning_rate": 1e-06,
"loss": 0.2754,
"step": 1751
},
{
"epoch": 0.669980879541109,
"grad_norm": 4.3634602536441,
"learning_rate": 1e-06,
"loss": 0.5403,
"step": 1752
},
{
"epoch": 0.6703632887189293,
"grad_norm": 2.234263063841419,
"learning_rate": 1e-06,
"loss": 0.4408,
"step": 1753
},
{
"epoch": 0.6707456978967495,
"grad_norm": 2.782095277044659,
"learning_rate": 1e-06,
"loss": 0.3659,
"step": 1754
},
{
"epoch": 0.6711281070745698,
"grad_norm": 3.3950645809183007,
"learning_rate": 1e-06,
"loss": 0.2794,
"step": 1755
},
{
"epoch": 0.6715105162523901,
"grad_norm": 3.319697783169732,
"learning_rate": 1e-06,
"loss": 0.4635,
"step": 1756
},
{
"epoch": 0.6718929254302103,
"grad_norm": 4.639040281423089,
"learning_rate": 1e-06,
"loss": 0.3632,
"step": 1757
},
{
"epoch": 0.6722753346080306,
"grad_norm": 2.7845421497937988,
"learning_rate": 1e-06,
"loss": 0.1857,
"step": 1758
},
{
"epoch": 0.6726577437858509,
"grad_norm": 4.246098017405284,
"learning_rate": 1e-06,
"loss": 0.2429,
"step": 1759
},
{
"epoch": 0.6730401529636711,
"grad_norm": 2.5775100436972322,
"learning_rate": 1e-06,
"loss": 0.1729,
"step": 1760
},
{
"epoch": 0.6734225621414914,
"grad_norm": 3.85421123307039,
"learning_rate": 1e-06,
"loss": 0.1487,
"step": 1761
},
{
"epoch": 0.6738049713193117,
"grad_norm": 4.20058843940997,
"learning_rate": 1e-06,
"loss": 0.1726,
"step": 1762
},
{
"epoch": 0.6741873804971319,
"grad_norm": 4.100446509261812,
"learning_rate": 1e-06,
"loss": 0.3263,
"step": 1763
},
{
"epoch": 0.6745697896749522,
"grad_norm": 3.6600236247816706,
"learning_rate": 1e-06,
"loss": 0.5095,
"step": 1764
},
{
"epoch": 0.6749521988527725,
"grad_norm": 3.121850372003473,
"learning_rate": 1e-06,
"loss": 0.371,
"step": 1765
},
{
"epoch": 0.6753346080305928,
"grad_norm": 2.2848251755027307,
"learning_rate": 1e-06,
"loss": 0.4956,
"step": 1766
},
{
"epoch": 0.675717017208413,
"grad_norm": 3.3260047140357583,
"learning_rate": 1e-06,
"loss": 0.4139,
"step": 1767
},
{
"epoch": 0.6760994263862332,
"grad_norm": 2.881912545840869,
"learning_rate": 1e-06,
"loss": 0.4155,
"step": 1768
},
{
"epoch": 0.6764818355640535,
"grad_norm": 3.464996231319255,
"learning_rate": 1e-06,
"loss": 0.2739,
"step": 1769
},
{
"epoch": 0.6768642447418738,
"grad_norm": 2.4101347837311367,
"learning_rate": 1e-06,
"loss": 0.2138,
"step": 1770
},
{
"epoch": 0.6772466539196941,
"grad_norm": 2.86264568108073,
"learning_rate": 1e-06,
"loss": 0.105,
"step": 1771
},
{
"epoch": 0.6776290630975144,
"grad_norm": 2.4440868096800394,
"learning_rate": 1e-06,
"loss": 0.1686,
"step": 1772
},
{
"epoch": 0.6780114722753346,
"grad_norm": 2.3747362693119,
"learning_rate": 1e-06,
"loss": 0.0947,
"step": 1773
},
{
"epoch": 0.6783938814531548,
"grad_norm": 7.99965380873262,
"learning_rate": 1e-06,
"loss": 0.222,
"step": 1774
},
{
"epoch": 0.6787762906309751,
"grad_norm": 4.487624635798562,
"learning_rate": 1e-06,
"loss": 0.2057,
"step": 1775
},
{
"epoch": 0.6791586998087954,
"grad_norm": 4.712107853042115,
"learning_rate": 1e-06,
"loss": 0.4001,
"step": 1776
},
{
"epoch": 0.6795411089866157,
"grad_norm": 2.910476728368903,
"learning_rate": 1e-06,
"loss": 0.6801,
"step": 1777
},
{
"epoch": 0.679923518164436,
"grad_norm": 2.90719490687544,
"learning_rate": 1e-06,
"loss": 0.4021,
"step": 1778
},
{
"epoch": 0.6803059273422563,
"grad_norm": 2.956284382077828,
"learning_rate": 1e-06,
"loss": 0.4785,
"step": 1779
},
{
"epoch": 0.6806883365200764,
"grad_norm": 3.7123519611063123,
"learning_rate": 1e-06,
"loss": 0.4861,
"step": 1780
},
{
"epoch": 0.6810707456978967,
"grad_norm": 2.3835400080491,
"learning_rate": 1e-06,
"loss": 0.5126,
"step": 1781
},
{
"epoch": 0.681453154875717,
"grad_norm": 2.0118450828809378,
"learning_rate": 1e-06,
"loss": 0.3863,
"step": 1782
},
{
"epoch": 0.6818355640535373,
"grad_norm": 2.476015337358834,
"learning_rate": 1e-06,
"loss": 0.223,
"step": 1783
},
{
"epoch": 0.6822179732313576,
"grad_norm": 2.9900879827204383,
"learning_rate": 1e-06,
"loss": 0.2357,
"step": 1784
},
{
"epoch": 0.6826003824091779,
"grad_norm": 4.134742790811418,
"learning_rate": 1e-06,
"loss": 0.2346,
"step": 1785
},
{
"epoch": 0.682982791586998,
"grad_norm": 4.008352142945461,
"learning_rate": 1e-06,
"loss": 0.1502,
"step": 1786
},
{
"epoch": 0.6833652007648183,
"grad_norm": 3.5316711444715336,
"learning_rate": 1e-06,
"loss": 0.0743,
"step": 1787
},
{
"epoch": 0.6837476099426386,
"grad_norm": 4.468546189148733,
"learning_rate": 1e-06,
"loss": 0.2104,
"step": 1788
},
{
"epoch": 0.6841300191204589,
"grad_norm": 4.987051600900471,
"learning_rate": 1e-06,
"loss": 0.6941,
"step": 1789
},
{
"epoch": 0.6845124282982792,
"grad_norm": 3.0797028346190394,
"learning_rate": 1e-06,
"loss": 0.508,
"step": 1790
},
{
"epoch": 0.6848948374760995,
"grad_norm": 3.0170605656983893,
"learning_rate": 1e-06,
"loss": 0.4026,
"step": 1791
},
{
"epoch": 0.6852772466539196,
"grad_norm": 3.1333803342445936,
"learning_rate": 1e-06,
"loss": 0.3969,
"step": 1792
},
{
"epoch": 0.6856596558317399,
"grad_norm": 2.9137190138234184,
"learning_rate": 1e-06,
"loss": 0.363,
"step": 1793
},
{
"epoch": 0.6860420650095602,
"grad_norm": 3.12110795840203,
"learning_rate": 1e-06,
"loss": 0.3825,
"step": 1794
},
{
"epoch": 0.6864244741873805,
"grad_norm": 2.0927769691626175,
"learning_rate": 1e-06,
"loss": 0.2357,
"step": 1795
},
{
"epoch": 0.6868068833652008,
"grad_norm": 3.637968870796287,
"learning_rate": 1e-06,
"loss": 0.3408,
"step": 1796
},
{
"epoch": 0.6871892925430211,
"grad_norm": 2.7240100707373522,
"learning_rate": 1e-06,
"loss": 0.222,
"step": 1797
},
{
"epoch": 0.6875717017208413,
"grad_norm": 4.31605781080897,
"learning_rate": 1e-06,
"loss": 0.1569,
"step": 1798
},
{
"epoch": 0.6879541108986615,
"grad_norm": 4.558239763730331,
"learning_rate": 1e-06,
"loss": 0.1587,
"step": 1799
},
{
"epoch": 0.6883365200764818,
"grad_norm": 4.345872538832599,
"learning_rate": 1e-06,
"loss": 0.1445,
"step": 1800
},
{
"epoch": 0.6887189292543021,
"grad_norm": 5.634707064795355,
"learning_rate": 1e-06,
"loss": 0.4121,
"step": 1801
},
{
"epoch": 0.6891013384321224,
"grad_norm": 3.095011906516201,
"learning_rate": 1e-06,
"loss": 0.471,
"step": 1802
},
{
"epoch": 0.6894837476099427,
"grad_norm": 6.63452931394001,
"learning_rate": 1e-06,
"loss": 0.4599,
"step": 1803
},
{
"epoch": 0.6898661567877629,
"grad_norm": 3.9377954765746828,
"learning_rate": 1e-06,
"loss": 0.2971,
"step": 1804
},
{
"epoch": 0.6902485659655831,
"grad_norm": 2.9167459931258204,
"learning_rate": 1e-06,
"loss": 0.546,
"step": 1805
},
{
"epoch": 0.6906309751434034,
"grad_norm": 3.3406928516773355,
"learning_rate": 1e-06,
"loss": 0.5512,
"step": 1806
},
{
"epoch": 0.6910133843212237,
"grad_norm": 3.420617138020482,
"learning_rate": 1e-06,
"loss": 0.2173,
"step": 1807
},
{
"epoch": 0.691395793499044,
"grad_norm": 2.6627656795135697,
"learning_rate": 1e-06,
"loss": 0.2218,
"step": 1808
},
{
"epoch": 0.6917782026768643,
"grad_norm": 3.961977489209145,
"learning_rate": 1e-06,
"loss": 0.1896,
"step": 1809
},
{
"epoch": 0.6921606118546845,
"grad_norm": 3.6100959883421306,
"learning_rate": 1e-06,
"loss": 0.0963,
"step": 1810
},
{
"epoch": 0.6925430210325048,
"grad_norm": 5.676831982382531,
"learning_rate": 1e-06,
"loss": 0.2514,
"step": 1811
},
{
"epoch": 0.692925430210325,
"grad_norm": 4.603937595014096,
"learning_rate": 1e-06,
"loss": 0.154,
"step": 1812
},
{
"epoch": 0.6933078393881453,
"grad_norm": 6.680118372890564,
"learning_rate": 1e-06,
"loss": 0.1824,
"step": 1813
},
{
"epoch": 0.6936902485659656,
"grad_norm": 4.771502762960885,
"learning_rate": 1e-06,
"loss": 0.75,
"step": 1814
},
{
"epoch": 0.6940726577437859,
"grad_norm": 3.6447441145215334,
"learning_rate": 1e-06,
"loss": 0.6686,
"step": 1815
},
{
"epoch": 0.6944550669216061,
"grad_norm": 2.5821162915328326,
"learning_rate": 1e-06,
"loss": 0.3767,
"step": 1816
},
{
"epoch": 0.6948374760994264,
"grad_norm": 3.5827351152104057,
"learning_rate": 1e-06,
"loss": 0.409,
"step": 1817
},
{
"epoch": 0.6952198852772467,
"grad_norm": 2.1990456114892005,
"learning_rate": 1e-06,
"loss": 0.3815,
"step": 1818
},
{
"epoch": 0.6956022944550669,
"grad_norm": 3.148419780066062,
"learning_rate": 1e-06,
"loss": 0.2984,
"step": 1819
},
{
"epoch": 0.6959847036328872,
"grad_norm": 3.385631628792315,
"learning_rate": 1e-06,
"loss": 0.2934,
"step": 1820
},
{
"epoch": 0.6963671128107075,
"grad_norm": 3.236165016682208,
"learning_rate": 1e-06,
"loss": 0.2187,
"step": 1821
},
{
"epoch": 0.6967495219885277,
"grad_norm": 2.7217993461877756,
"learning_rate": 1e-06,
"loss": 0.1829,
"step": 1822
},
{
"epoch": 0.697131931166348,
"grad_norm": 4.96599324843293,
"learning_rate": 1e-06,
"loss": 0.219,
"step": 1823
},
{
"epoch": 0.6975143403441683,
"grad_norm": 3.333908428808765,
"learning_rate": 1e-06,
"loss": 0.0983,
"step": 1824
},
{
"epoch": 0.6978967495219885,
"grad_norm": 5.570685033124853,
"learning_rate": 1e-06,
"loss": 0.263,
"step": 1825
},
{
"epoch": 0.6982791586998088,
"grad_norm": 6.411933573492392,
"learning_rate": 1e-06,
"loss": 0.2969,
"step": 1826
},
{
"epoch": 0.698661567877629,
"grad_norm": 3.775101636314174,
"learning_rate": 1e-06,
"loss": 0.5427,
"step": 1827
},
{
"epoch": 0.6990439770554493,
"grad_norm": 7.071409024518453,
"learning_rate": 1e-06,
"loss": 0.6173,
"step": 1828
},
{
"epoch": 0.6994263862332696,
"grad_norm": 3.762640439070742,
"learning_rate": 1e-06,
"loss": 0.3903,
"step": 1829
},
{
"epoch": 0.6998087954110899,
"grad_norm": 3.2679992513854197,
"learning_rate": 1e-06,
"loss": 0.3204,
"step": 1830
},
{
"epoch": 0.7001912045889102,
"grad_norm": 3.7052534079024153,
"learning_rate": 1e-06,
"loss": 0.577,
"step": 1831
},
{
"epoch": 0.7005736137667304,
"grad_norm": 2.291632033577672,
"learning_rate": 1e-06,
"loss": 0.3083,
"step": 1832
},
{
"epoch": 0.7009560229445506,
"grad_norm": 3.1704137283039517,
"learning_rate": 1e-06,
"loss": 0.2115,
"step": 1833
},
{
"epoch": 0.7013384321223709,
"grad_norm": 2.7626738238990645,
"learning_rate": 1e-06,
"loss": 0.2287,
"step": 1834
},
{
"epoch": 0.7017208413001912,
"grad_norm": 3.2028281469723106,
"learning_rate": 1e-06,
"loss": 0.2692,
"step": 1835
},
{
"epoch": 0.7021032504780115,
"grad_norm": 3.9638971410776977,
"learning_rate": 1e-06,
"loss": 0.121,
"step": 1836
},
{
"epoch": 0.7024856596558318,
"grad_norm": 2.8519179867245357,
"learning_rate": 1e-06,
"loss": 0.0548,
"step": 1837
},
{
"epoch": 0.702868068833652,
"grad_norm": 4.584365497311428,
"learning_rate": 1e-06,
"loss": 0.2794,
"step": 1838
},
{
"epoch": 0.7032504780114722,
"grad_norm": 5.193457985207564,
"learning_rate": 1e-06,
"loss": 0.4874,
"step": 1839
},
{
"epoch": 0.7036328871892925,
"grad_norm": 3.693068744555693,
"learning_rate": 1e-06,
"loss": 0.6087,
"step": 1840
},
{
"epoch": 0.7040152963671128,
"grad_norm": 2.6873466093069327,
"learning_rate": 1e-06,
"loss": 0.4215,
"step": 1841
},
{
"epoch": 0.7043977055449331,
"grad_norm": 4.868728516204405,
"learning_rate": 1e-06,
"loss": 0.4795,
"step": 1842
},
{
"epoch": 0.7047801147227534,
"grad_norm": 2.704295087056366,
"learning_rate": 1e-06,
"loss": 0.3664,
"step": 1843
},
{
"epoch": 0.7051625239005737,
"grad_norm": 2.0360407746734377,
"learning_rate": 1e-06,
"loss": 0.181,
"step": 1844
},
{
"epoch": 0.7055449330783938,
"grad_norm": 2.5505686929886457,
"learning_rate": 1e-06,
"loss": 0.1876,
"step": 1845
},
{
"epoch": 0.7059273422562141,
"grad_norm": 2.925389588323107,
"learning_rate": 1e-06,
"loss": 0.2242,
"step": 1846
},
{
"epoch": 0.7063097514340344,
"grad_norm": 3.9714455649276004,
"learning_rate": 1e-06,
"loss": 0.2156,
"step": 1847
},
{
"epoch": 0.7066921606118547,
"grad_norm": 2.8432737517589484,
"learning_rate": 1e-06,
"loss": 0.1292,
"step": 1848
},
{
"epoch": 0.707074569789675,
"grad_norm": 3.3187594023892824,
"learning_rate": 1e-06,
"loss": 0.0879,
"step": 1849
},
{
"epoch": 0.7074569789674953,
"grad_norm": 4.174831071736803,
"learning_rate": 1e-06,
"loss": 0.1524,
"step": 1850
},
{
"epoch": 0.7078393881453154,
"grad_norm": 4.211490516939054,
"learning_rate": 1e-06,
"loss": 0.3833,
"step": 1851
},
{
"epoch": 0.7082217973231357,
"grad_norm": 4.472239592355886,
"learning_rate": 1e-06,
"loss": 0.7896,
"step": 1852
},
{
"epoch": 0.708604206500956,
"grad_norm": 3.334509800882675,
"learning_rate": 1e-06,
"loss": 0.4439,
"step": 1853
},
{
"epoch": 0.7089866156787763,
"grad_norm": 2.205621049401282,
"learning_rate": 1e-06,
"loss": 0.5861,
"step": 1854
},
{
"epoch": 0.7093690248565966,
"grad_norm": 3.6115895533078644,
"learning_rate": 1e-06,
"loss": 0.5834,
"step": 1855
},
{
"epoch": 0.7097514340344169,
"grad_norm": 2.5330680628810085,
"learning_rate": 1e-06,
"loss": 0.41,
"step": 1856
},
{
"epoch": 0.710133843212237,
"grad_norm": 1.9168456104613885,
"learning_rate": 1e-06,
"loss": 0.3146,
"step": 1857
},
{
"epoch": 0.7105162523900573,
"grad_norm": 3.2846287361661197,
"learning_rate": 1e-06,
"loss": 0.3478,
"step": 1858
},
{
"epoch": 0.7108986615678776,
"grad_norm": 2.62089053416537,
"learning_rate": 1e-06,
"loss": 0.1584,
"step": 1859
},
{
"epoch": 0.7112810707456979,
"grad_norm": 4.912714301717813,
"learning_rate": 1e-06,
"loss": 0.2955,
"step": 1860
},
{
"epoch": 0.7116634799235182,
"grad_norm": 3.5978899122935775,
"learning_rate": 1e-06,
"loss": 0.0844,
"step": 1861
},
{
"epoch": 0.7120458891013385,
"grad_norm": 4.862546027193691,
"learning_rate": 1e-06,
"loss": 0.1187,
"step": 1862
},
{
"epoch": 0.7124282982791587,
"grad_norm": 6.103384061082042,
"learning_rate": 1e-06,
"loss": 0.1864,
"step": 1863
},
{
"epoch": 0.7128107074569789,
"grad_norm": 4.6811541518214534,
"learning_rate": 1e-06,
"loss": 0.565,
"step": 1864
},
{
"epoch": 0.7131931166347992,
"grad_norm": 3.4928539075510487,
"learning_rate": 1e-06,
"loss": 0.5202,
"step": 1865
},
{
"epoch": 0.7135755258126195,
"grad_norm": 3.4404708509015625,
"learning_rate": 1e-06,
"loss": 0.3804,
"step": 1866
},
{
"epoch": 0.7139579349904398,
"grad_norm": 2.1495928622815477,
"learning_rate": 1e-06,
"loss": 0.5346,
"step": 1867
},
{
"epoch": 0.7143403441682601,
"grad_norm": 2.6974655618670087,
"learning_rate": 1e-06,
"loss": 0.505,
"step": 1868
},
{
"epoch": 0.7147227533460803,
"grad_norm": 4.0097496898276885,
"learning_rate": 1e-06,
"loss": 0.3809,
"step": 1869
},
{
"epoch": 0.7151051625239006,
"grad_norm": 3.45407519493576,
"learning_rate": 1e-06,
"loss": 0.3726,
"step": 1870
},
{
"epoch": 0.7154875717017208,
"grad_norm": 3.2900777595397894,
"learning_rate": 1e-06,
"loss": 0.2127,
"step": 1871
},
{
"epoch": 0.7158699808795411,
"grad_norm": 4.368430736920466,
"learning_rate": 1e-06,
"loss": 0.1821,
"step": 1872
},
{
"epoch": 0.7162523900573614,
"grad_norm": 3.4827469242403564,
"learning_rate": 1e-06,
"loss": 0.2243,
"step": 1873
},
{
"epoch": 0.7166347992351817,
"grad_norm": 4.181869137118051,
"learning_rate": 1e-06,
"loss": 0.1036,
"step": 1874
},
{
"epoch": 0.7170172084130019,
"grad_norm": 6.611198300054725,
"learning_rate": 1e-06,
"loss": 0.2067,
"step": 1875
},
{
"epoch": 0.7173996175908222,
"grad_norm": 6.57649236067522,
"learning_rate": 1e-06,
"loss": 0.6051,
"step": 1876
},
{
"epoch": 0.7177820267686424,
"grad_norm": 4.121052326880088,
"learning_rate": 1e-06,
"loss": 0.6721,
"step": 1877
},
{
"epoch": 0.7181644359464627,
"grad_norm": 3.35239881916105,
"learning_rate": 1e-06,
"loss": 0.5213,
"step": 1878
},
{
"epoch": 0.718546845124283,
"grad_norm": 3.1891759318654165,
"learning_rate": 1e-06,
"loss": 0.3168,
"step": 1879
},
{
"epoch": 0.7189292543021033,
"grad_norm": 3.4945218948112817,
"learning_rate": 1e-06,
"loss": 0.2938,
"step": 1880
},
{
"epoch": 0.7193116634799235,
"grad_norm": 3.5559991784969105,
"learning_rate": 1e-06,
"loss": 0.2688,
"step": 1881
},
{
"epoch": 0.7196940726577438,
"grad_norm": 2.812100784790976,
"learning_rate": 1e-06,
"loss": 0.3084,
"step": 1882
},
{
"epoch": 0.720076481835564,
"grad_norm": 2.4886104539636467,
"learning_rate": 1e-06,
"loss": 0.1731,
"step": 1883
},
{
"epoch": 0.7204588910133843,
"grad_norm": 3.2822932446830215,
"learning_rate": 1e-06,
"loss": 0.1773,
"step": 1884
},
{
"epoch": 0.7208413001912046,
"grad_norm": 2.9886354722655595,
"learning_rate": 1e-06,
"loss": 0.1471,
"step": 1885
},
{
"epoch": 0.7212237093690248,
"grad_norm": 4.083413389290498,
"learning_rate": 1e-06,
"loss": 0.162,
"step": 1886
},
{
"epoch": 0.7216061185468451,
"grad_norm": 3.2348496830492492,
"learning_rate": 1e-06,
"loss": 0.1084,
"step": 1887
},
{
"epoch": 0.7219885277246654,
"grad_norm": 6.62174994443766,
"learning_rate": 1e-06,
"loss": 0.3879,
"step": 1888
},
{
"epoch": 0.7223709369024857,
"grad_norm": 4.169378212144267,
"learning_rate": 1e-06,
"loss": 0.5458,
"step": 1889
},
{
"epoch": 0.722753346080306,
"grad_norm": 4.000780506279222,
"learning_rate": 1e-06,
"loss": 0.3392,
"step": 1890
},
{
"epoch": 0.7231357552581262,
"grad_norm": 2.150841162498716,
"learning_rate": 1e-06,
"loss": 0.1693,
"step": 1891
},
{
"epoch": 0.7235181644359464,
"grad_norm": 2.410613030623257,
"learning_rate": 1e-06,
"loss": 0.5151,
"step": 1892
},
{
"epoch": 0.7239005736137667,
"grad_norm": 2.9877919712803735,
"learning_rate": 1e-06,
"loss": 0.5863,
"step": 1893
},
{
"epoch": 0.724282982791587,
"grad_norm": 4.180110503565843,
"learning_rate": 1e-06,
"loss": 0.4053,
"step": 1894
},
{
"epoch": 0.7246653919694073,
"grad_norm": 3.0684353669078828,
"learning_rate": 1e-06,
"loss": 0.4632,
"step": 1895
},
{
"epoch": 0.7250478011472276,
"grad_norm": 3.7202368576774356,
"learning_rate": 1e-06,
"loss": 0.2632,
"step": 1896
},
{
"epoch": 0.7254302103250478,
"grad_norm": 2.7200966492479637,
"learning_rate": 1e-06,
"loss": 0.2772,
"step": 1897
},
{
"epoch": 0.725812619502868,
"grad_norm": 2.4582363260386018,
"learning_rate": 1e-06,
"loss": 0.1286,
"step": 1898
},
{
"epoch": 0.7261950286806883,
"grad_norm": 4.900075927457515,
"learning_rate": 1e-06,
"loss": 0.1468,
"step": 1899
},
{
"epoch": 0.7265774378585086,
"grad_norm": 5.773966722403739,
"learning_rate": 1e-06,
"loss": 0.1803,
"step": 1900
},
{
"epoch": 0.7269598470363289,
"grad_norm": 4.439438517353882,
"learning_rate": 1e-06,
"loss": 0.4792,
"step": 1901
},
{
"epoch": 0.7273422562141492,
"grad_norm": 3.5652131574126233,
"learning_rate": 1e-06,
"loss": 0.5077,
"step": 1902
},
{
"epoch": 0.7277246653919694,
"grad_norm": 4.41994129400838,
"learning_rate": 1e-06,
"loss": 0.4999,
"step": 1903
},
{
"epoch": 0.7281070745697896,
"grad_norm": 2.0358063287410446,
"learning_rate": 1e-06,
"loss": 0.5988,
"step": 1904
},
{
"epoch": 0.7284894837476099,
"grad_norm": 2.6470563770108355,
"learning_rate": 1e-06,
"loss": 0.3909,
"step": 1905
},
{
"epoch": 0.7288718929254302,
"grad_norm": 3.8597156204460625,
"learning_rate": 1e-06,
"loss": 0.3542,
"step": 1906
},
{
"epoch": 0.7292543021032505,
"grad_norm": 3.1934985395899718,
"learning_rate": 1e-06,
"loss": 0.2437,
"step": 1907
},
{
"epoch": 0.7296367112810708,
"grad_norm": 3.2701656040639486,
"learning_rate": 1e-06,
"loss": 0.1474,
"step": 1908
},
{
"epoch": 0.7300191204588911,
"grad_norm": 2.4833295049787196,
"learning_rate": 1e-06,
"loss": 0.2791,
"step": 1909
},
{
"epoch": 0.7304015296367112,
"grad_norm": 3.7614979268161752,
"learning_rate": 1e-06,
"loss": 0.1431,
"step": 1910
},
{
"epoch": 0.7307839388145315,
"grad_norm": 4.869833139933135,
"learning_rate": 1e-06,
"loss": 0.2185,
"step": 1911
},
{
"epoch": 0.7311663479923518,
"grad_norm": 4.298374649525742,
"learning_rate": 1e-06,
"loss": 0.1067,
"step": 1912
},
{
"epoch": 0.7315487571701721,
"grad_norm": 6.476052700174656,
"learning_rate": 1e-06,
"loss": 0.3223,
"step": 1913
},
{
"epoch": 0.7319311663479924,
"grad_norm": 4.323621920618311,
"learning_rate": 1e-06,
"loss": 0.3189,
"step": 1914
},
{
"epoch": 0.7323135755258127,
"grad_norm": 2.657115211943629,
"learning_rate": 1e-06,
"loss": 0.3768,
"step": 1915
},
{
"epoch": 0.7326959847036328,
"grad_norm": 3.37992019875126,
"learning_rate": 1e-06,
"loss": 0.3918,
"step": 1916
},
{
"epoch": 0.7330783938814531,
"grad_norm": 3.1789435938762542,
"learning_rate": 1e-06,
"loss": 0.3355,
"step": 1917
},
{
"epoch": 0.7334608030592734,
"grad_norm": 3.519897035479995,
"learning_rate": 1e-06,
"loss": 0.3091,
"step": 1918
},
{
"epoch": 0.7338432122370937,
"grad_norm": 2.9668340021438433,
"learning_rate": 1e-06,
"loss": 0.3655,
"step": 1919
},
{
"epoch": 0.734225621414914,
"grad_norm": 4.146447115500284,
"learning_rate": 1e-06,
"loss": 0.2495,
"step": 1920
},
{
"epoch": 0.7346080305927343,
"grad_norm": 2.6198207077530102,
"learning_rate": 1e-06,
"loss": 0.1357,
"step": 1921
},
{
"epoch": 0.7349904397705544,
"grad_norm": 3.949068664657057,
"learning_rate": 1e-06,
"loss": 0.2524,
"step": 1922
},
{
"epoch": 0.7353728489483747,
"grad_norm": 4.445273669028006,
"learning_rate": 1e-06,
"loss": 0.152,
"step": 1923
},
{
"epoch": 0.735755258126195,
"grad_norm": 3.1655209794422565,
"learning_rate": 1e-06,
"loss": 0.1172,
"step": 1924
},
{
"epoch": 0.7361376673040153,
"grad_norm": 3.7767993806941504,
"learning_rate": 1e-06,
"loss": 0.1635,
"step": 1925
},
{
"epoch": 0.7365200764818356,
"grad_norm": 4.993509848767819,
"learning_rate": 1e-06,
"loss": 0.2869,
"step": 1926
},
{
"epoch": 0.7369024856596559,
"grad_norm": 3.545566401462217,
"learning_rate": 1e-06,
"loss": 0.5885,
"step": 1927
},
{
"epoch": 0.7372848948374761,
"grad_norm": 10.560810382843867,
"learning_rate": 1e-06,
"loss": 0.4926,
"step": 1928
},
{
"epoch": 0.7376673040152963,
"grad_norm": 2.330562058466115,
"learning_rate": 1e-06,
"loss": 0.4563,
"step": 1929
},
{
"epoch": 0.7380497131931166,
"grad_norm": 2.144996022929882,
"learning_rate": 1e-06,
"loss": 0.314,
"step": 1930
},
{
"epoch": 0.7384321223709369,
"grad_norm": 3.0278741974794325,
"learning_rate": 1e-06,
"loss": 0.3752,
"step": 1931
},
{
"epoch": 0.7388145315487572,
"grad_norm": 2.002602790927952,
"learning_rate": 1e-06,
"loss": 0.3564,
"step": 1932
},
{
"epoch": 0.7391969407265775,
"grad_norm": 3.743961558788568,
"learning_rate": 1e-06,
"loss": 0.399,
"step": 1933
},
{
"epoch": 0.7395793499043977,
"grad_norm": 3.8753553658246336,
"learning_rate": 1e-06,
"loss": 0.3506,
"step": 1934
},
{
"epoch": 0.739961759082218,
"grad_norm": 3.002976847952848,
"learning_rate": 1e-06,
"loss": 0.2142,
"step": 1935
},
{
"epoch": 0.7403441682600382,
"grad_norm": 2.0153218122643652,
"learning_rate": 1e-06,
"loss": 0.057,
"step": 1936
},
{
"epoch": 0.7407265774378585,
"grad_norm": 3.0885672265335886,
"learning_rate": 1e-06,
"loss": 0.0932,
"step": 1937
},
{
"epoch": 0.7411089866156788,
"grad_norm": 4.183245419088883,
"learning_rate": 1e-06,
"loss": 0.2852,
"step": 1938
},
{
"epoch": 0.7414913957934991,
"grad_norm": 4.42740417514626,
"learning_rate": 1e-06,
"loss": 0.556,
"step": 1939
},
{
"epoch": 0.7418738049713193,
"grad_norm": 3.603905699585369,
"learning_rate": 1e-06,
"loss": 0.4392,
"step": 1940
},
{
"epoch": 0.7422562141491396,
"grad_norm": 3.007430411506426,
"learning_rate": 1e-06,
"loss": 0.3079,
"step": 1941
},
{
"epoch": 0.7426386233269598,
"grad_norm": 1.5184469006913335,
"learning_rate": 1e-06,
"loss": 0.284,
"step": 1942
},
{
"epoch": 0.7430210325047801,
"grad_norm": 2.088187044118794,
"learning_rate": 1e-06,
"loss": 0.3736,
"step": 1943
},
{
"epoch": 0.7434034416826004,
"grad_norm": 3.5731852053310473,
"learning_rate": 1e-06,
"loss": 0.5477,
"step": 1944
},
{
"epoch": 0.7437858508604207,
"grad_norm": 2.961146525953916,
"learning_rate": 1e-06,
"loss": 0.3065,
"step": 1945
},
{
"epoch": 0.7441682600382409,
"grad_norm": 3.6100323231272373,
"learning_rate": 1e-06,
"loss": 0.2124,
"step": 1946
},
{
"epoch": 0.7445506692160612,
"grad_norm": 2.7169788774070143,
"learning_rate": 1e-06,
"loss": 0.252,
"step": 1947
},
{
"epoch": 0.7449330783938815,
"grad_norm": 4.019746201887364,
"learning_rate": 1e-06,
"loss": 0.2892,
"step": 1948
},
{
"epoch": 0.7453154875717017,
"grad_norm": 3.0840721963429876,
"learning_rate": 1e-06,
"loss": 0.1068,
"step": 1949
},
{
"epoch": 0.745697896749522,
"grad_norm": 4.265207347055815,
"learning_rate": 1e-06,
"loss": 0.2368,
"step": 1950
},
{
"epoch": 0.7460803059273422,
"grad_norm": 5.169757882209663,
"learning_rate": 1e-06,
"loss": 0.4368,
"step": 1951
},
{
"epoch": 0.7464627151051625,
"grad_norm": 3.5514206850845325,
"learning_rate": 1e-06,
"loss": 0.5124,
"step": 1952
},
{
"epoch": 0.7468451242829828,
"grad_norm": 3.3868851670293685,
"learning_rate": 1e-06,
"loss": 0.3539,
"step": 1953
},
{
"epoch": 0.7472275334608031,
"grad_norm": 3.3671307127516408,
"learning_rate": 1e-06,
"loss": 0.3289,
"step": 1954
},
{
"epoch": 0.7476099426386233,
"grad_norm": 3.7899307357377197,
"learning_rate": 1e-06,
"loss": 0.4966,
"step": 1955
},
{
"epoch": 0.7479923518164436,
"grad_norm": 2.3303891635442464,
"learning_rate": 1e-06,
"loss": 0.2466,
"step": 1956
},
{
"epoch": 0.7483747609942638,
"grad_norm": 4.211143587692316,
"learning_rate": 1e-06,
"loss": 0.468,
"step": 1957
},
{
"epoch": 0.7487571701720841,
"grad_norm": 2.4047094590719373,
"learning_rate": 1e-06,
"loss": 0.2326,
"step": 1958
},
{
"epoch": 0.7491395793499044,
"grad_norm": 3.3531454834041328,
"learning_rate": 1e-06,
"loss": 0.3103,
"step": 1959
},
{
"epoch": 0.7495219885277247,
"grad_norm": 3.2136104994727934,
"learning_rate": 1e-06,
"loss": 0.2424,
"step": 1960
},
{
"epoch": 0.749904397705545,
"grad_norm": 4.63095925431355,
"learning_rate": 1e-06,
"loss": 0.1649,
"step": 1961
},
{
"epoch": 0.7502868068833652,
"grad_norm": 4.552692899183663,
"learning_rate": 1e-06,
"loss": 0.1504,
"step": 1962
},
{
"epoch": 0.7506692160611854,
"grad_norm": 6.039571289106634,
"learning_rate": 1e-06,
"loss": 0.3453,
"step": 1963
},
{
"epoch": 0.7510516252390057,
"grad_norm": 6.0466352706032405,
"learning_rate": 1e-06,
"loss": 0.5867,
"step": 1964
},
{
"epoch": 0.751434034416826,
"grad_norm": 2.1601803107413398,
"learning_rate": 1e-06,
"loss": 0.3917,
"step": 1965
},
{
"epoch": 0.7518164435946463,
"grad_norm": 1.9370251965949365,
"learning_rate": 1e-06,
"loss": 0.1983,
"step": 1966
},
{
"epoch": 0.7521988527724666,
"grad_norm": 2.6129106335779917,
"learning_rate": 1e-06,
"loss": 0.1441,
"step": 1967
},
{
"epoch": 0.7525812619502869,
"grad_norm": 4.027613455975739,
"learning_rate": 1e-06,
"loss": 0.4051,
"step": 1968
},
{
"epoch": 0.752963671128107,
"grad_norm": 3.8408939326388887,
"learning_rate": 1e-06,
"loss": 0.4728,
"step": 1969
},
{
"epoch": 0.7533460803059273,
"grad_norm": 3.506971908478582,
"learning_rate": 1e-06,
"loss": 0.3481,
"step": 1970
},
{
"epoch": 0.7537284894837476,
"grad_norm": 3.4824632266719013,
"learning_rate": 1e-06,
"loss": 0.3397,
"step": 1971
},
{
"epoch": 0.7541108986615679,
"grad_norm": 3.8161371110799367,
"learning_rate": 1e-06,
"loss": 0.2475,
"step": 1972
},
{
"epoch": 0.7544933078393882,
"grad_norm": 3.2389161300894944,
"learning_rate": 1e-06,
"loss": 0.1609,
"step": 1973
},
{
"epoch": 0.7548757170172085,
"grad_norm": 3.676404650826159,
"learning_rate": 1e-06,
"loss": 0.1427,
"step": 1974
},
{
"epoch": 0.7552581261950286,
"grad_norm": 4.308149492976423,
"learning_rate": 1e-06,
"loss": 0.1601,
"step": 1975
},
{
"epoch": 0.7556405353728489,
"grad_norm": 4.053353915371422,
"learning_rate": 1e-06,
"loss": 0.5546,
"step": 1976
},
{
"epoch": 0.7560229445506692,
"grad_norm": 3.8180897493119867,
"learning_rate": 1e-06,
"loss": 0.4642,
"step": 1977
},
{
"epoch": 0.7564053537284895,
"grad_norm": 3.779026354982245,
"learning_rate": 1e-06,
"loss": 0.41,
"step": 1978
},
{
"epoch": 0.7567877629063098,
"grad_norm": 3.352605484021963,
"learning_rate": 1e-06,
"loss": 0.4069,
"step": 1979
},
{
"epoch": 0.7571701720841301,
"grad_norm": 2.5243564503217946,
"learning_rate": 1e-06,
"loss": 0.3045,
"step": 1980
},
{
"epoch": 0.7575525812619502,
"grad_norm": 3.1271164402913008,
"learning_rate": 1e-06,
"loss": 0.178,
"step": 1981
},
{
"epoch": 0.7579349904397705,
"grad_norm": 2.1511791145309562,
"learning_rate": 1e-06,
"loss": 0.281,
"step": 1982
},
{
"epoch": 0.7583173996175908,
"grad_norm": 3.538560709632089,
"learning_rate": 1e-06,
"loss": 0.3045,
"step": 1983
},
{
"epoch": 0.7586998087954111,
"grad_norm": 2.408642755802393,
"learning_rate": 1e-06,
"loss": 0.2651,
"step": 1984
},
{
"epoch": 0.7590822179732314,
"grad_norm": 5.043234631231656,
"learning_rate": 1e-06,
"loss": 0.295,
"step": 1985
},
{
"epoch": 0.7594646271510517,
"grad_norm": 2.754032299863494,
"learning_rate": 1e-06,
"loss": 0.0662,
"step": 1986
},
{
"epoch": 0.7598470363288718,
"grad_norm": 5.2173929791516915,
"learning_rate": 1e-06,
"loss": 0.1823,
"step": 1987
},
{
"epoch": 0.7602294455066921,
"grad_norm": 4.407626755352812,
"learning_rate": 1e-06,
"loss": 0.2591,
"step": 1988
},
{
"epoch": 0.7606118546845124,
"grad_norm": 3.548506575751531,
"learning_rate": 1e-06,
"loss": 0.4805,
"step": 1989
},
{
"epoch": 0.7609942638623327,
"grad_norm": 3.728761547964631,
"learning_rate": 1e-06,
"loss": 0.4266,
"step": 1990
},
{
"epoch": 0.761376673040153,
"grad_norm": 2.82515176686109,
"learning_rate": 1e-06,
"loss": 0.527,
"step": 1991
},
{
"epoch": 0.7617590822179733,
"grad_norm": 2.477642605985699,
"learning_rate": 1e-06,
"loss": 0.3361,
"step": 1992
},
{
"epoch": 0.7621414913957935,
"grad_norm": 2.260690827744086,
"learning_rate": 1e-06,
"loss": 0.4581,
"step": 1993
},
{
"epoch": 0.7625239005736137,
"grad_norm": 4.9252590595408146,
"learning_rate": 1e-06,
"loss": 0.2767,
"step": 1994
},
{
"epoch": 0.762906309751434,
"grad_norm": 2.867861967247307,
"learning_rate": 1e-06,
"loss": 0.1954,
"step": 1995
},
{
"epoch": 0.7632887189292543,
"grad_norm": 3.9892459789833667,
"learning_rate": 1e-06,
"loss": 0.173,
"step": 1996
},
{
"epoch": 0.7636711281070746,
"grad_norm": 2.8326691428688644,
"learning_rate": 1e-06,
"loss": 0.2761,
"step": 1997
},
{
"epoch": 0.7640535372848949,
"grad_norm": 5.023119309761381,
"learning_rate": 1e-06,
"loss": 0.2801,
"step": 1998
},
{
"epoch": 0.7644359464627151,
"grad_norm": 4.259385684497444,
"learning_rate": 1e-06,
"loss": 0.2594,
"step": 1999
},
{
"epoch": 0.7648183556405354,
"grad_norm": 4.156254359651731,
"learning_rate": 1e-06,
"loss": 0.1968,
"step": 2000
},
{
"epoch": 0.7648183556405354,
"eval_runtime": 805.9648,
"eval_samples_per_second": 1.903,
"eval_steps_per_second": 0.476,
"step": 2000
},
{
"epoch": 0.7652007648183556,
"grad_norm": 4.5302225987056355,
"learning_rate": 1e-06,
"loss": 0.4251,
"step": 2001
},
{
"epoch": 0.7655831739961759,
"grad_norm": 3.6469216311888584,
"learning_rate": 1e-06,
"loss": 0.5221,
"step": 2002
},
{
"epoch": 0.7659655831739962,
"grad_norm": 4.288490576907294,
"learning_rate": 1e-06,
"loss": 0.6978,
"step": 2003
},
{
"epoch": 0.7663479923518165,
"grad_norm": 2.6543820264349183,
"learning_rate": 1e-06,
"loss": 0.3381,
"step": 2004
},
{
"epoch": 0.7667304015296367,
"grad_norm": 1.9267132652921741,
"learning_rate": 1e-06,
"loss": 0.3503,
"step": 2005
},
{
"epoch": 0.767112810707457,
"grad_norm": 2.7465367617797094,
"learning_rate": 1e-06,
"loss": 0.366,
"step": 2006
},
{
"epoch": 0.7674952198852772,
"grad_norm": 5.328444306322995,
"learning_rate": 1e-06,
"loss": 0.49,
"step": 2007
},
{
"epoch": 0.7678776290630975,
"grad_norm": 3.017219398681553,
"learning_rate": 1e-06,
"loss": 0.2735,
"step": 2008
},
{
"epoch": 0.7682600382409178,
"grad_norm": 1.95948573625478,
"learning_rate": 1e-06,
"loss": 0.2887,
"step": 2009
},
{
"epoch": 0.768642447418738,
"grad_norm": 3.835460003814006,
"learning_rate": 1e-06,
"loss": 0.1644,
"step": 2010
},
{
"epoch": 0.7690248565965583,
"grad_norm": 2.9154317875633056,
"learning_rate": 1e-06,
"loss": 0.2051,
"step": 2011
},
{
"epoch": 0.7694072657743786,
"grad_norm": 3.836144342692119,
"learning_rate": 1e-06,
"loss": 0.1727,
"step": 2012
},
{
"epoch": 0.7697896749521989,
"grad_norm": 4.068723635727276,
"learning_rate": 1e-06,
"loss": 0.2228,
"step": 2013
},
{
"epoch": 0.7701720841300191,
"grad_norm": 5.441157021055651,
"learning_rate": 1e-06,
"loss": 0.8692,
"step": 2014
},
{
"epoch": 0.7705544933078394,
"grad_norm": 3.608137438759441,
"learning_rate": 1e-06,
"loss": 0.4858,
"step": 2015
},
{
"epoch": 0.7709369024856596,
"grad_norm": 1.690505283368894,
"learning_rate": 1e-06,
"loss": 0.4245,
"step": 2016
},
{
"epoch": 0.7713193116634799,
"grad_norm": 2.8006470034892073,
"learning_rate": 1e-06,
"loss": 0.3648,
"step": 2017
},
{
"epoch": 0.7717017208413002,
"grad_norm": 3.9321450044005655,
"learning_rate": 1e-06,
"loss": 0.4713,
"step": 2018
},
{
"epoch": 0.7720841300191205,
"grad_norm": 3.3301640384811915,
"learning_rate": 1e-06,
"loss": 0.2594,
"step": 2019
},
{
"epoch": 0.7724665391969407,
"grad_norm": 2.8530852967891436,
"learning_rate": 1e-06,
"loss": 0.4297,
"step": 2020
},
{
"epoch": 0.772848948374761,
"grad_norm": 3.9360004778915982,
"learning_rate": 1e-06,
"loss": 0.3659,
"step": 2021
},
{
"epoch": 0.7732313575525812,
"grad_norm": 3.245990333791879,
"learning_rate": 1e-06,
"loss": 0.2666,
"step": 2022
},
{
"epoch": 0.7736137667304015,
"grad_norm": 3.5073750314453194,
"learning_rate": 1e-06,
"loss": 0.2183,
"step": 2023
},
{
"epoch": 0.7739961759082218,
"grad_norm": 2.669386390727826,
"learning_rate": 1e-06,
"loss": 0.1312,
"step": 2024
},
{
"epoch": 0.7743785850860421,
"grad_norm": 4.800203144225261,
"learning_rate": 1e-06,
"loss": 0.2841,
"step": 2025
},
{
"epoch": 0.7747609942638624,
"grad_norm": 5.814659783865227,
"learning_rate": 1e-06,
"loss": 0.3287,
"step": 2026
},
{
"epoch": 0.7751434034416826,
"grad_norm": 3.2658903023307304,
"learning_rate": 1e-06,
"loss": 0.4673,
"step": 2027
},
{
"epoch": 0.7755258126195028,
"grad_norm": 3.1296906934503803,
"learning_rate": 1e-06,
"loss": 0.5212,
"step": 2028
},
{
"epoch": 0.7759082217973231,
"grad_norm": 2.535589195865614,
"learning_rate": 1e-06,
"loss": 0.4094,
"step": 2029
},
{
"epoch": 0.7762906309751434,
"grad_norm": 2.3952707265855557,
"learning_rate": 1e-06,
"loss": 0.2946,
"step": 2030
},
{
"epoch": 0.7766730401529637,
"grad_norm": 3.7295965163628253,
"learning_rate": 1e-06,
"loss": 0.3445,
"step": 2031
},
{
"epoch": 0.777055449330784,
"grad_norm": 2.110488597651997,
"learning_rate": 1e-06,
"loss": 0.2046,
"step": 2032
},
{
"epoch": 0.7774378585086043,
"grad_norm": 2.665524078998848,
"learning_rate": 1e-06,
"loss": 0.1469,
"step": 2033
},
{
"epoch": 0.7778202676864244,
"grad_norm": 3.196103620642146,
"learning_rate": 1e-06,
"loss": 0.2041,
"step": 2034
},
{
"epoch": 0.7782026768642447,
"grad_norm": 5.1756694047467375,
"learning_rate": 1e-06,
"loss": 0.2137,
"step": 2035
},
{
"epoch": 0.778585086042065,
"grad_norm": 5.961311538580228,
"learning_rate": 1e-06,
"loss": 0.1228,
"step": 2036
},
{
"epoch": 0.7789674952198853,
"grad_norm": 4.2007380518253825,
"learning_rate": 1e-06,
"loss": 0.0702,
"step": 2037
},
{
"epoch": 0.7793499043977056,
"grad_norm": 6.277067243637935,
"learning_rate": 1e-06,
"loss": 0.1584,
"step": 2038
},
{
"epoch": 0.7797323135755259,
"grad_norm": 3.8274154316873874,
"learning_rate": 1e-06,
"loss": 0.666,
"step": 2039
},
{
"epoch": 0.780114722753346,
"grad_norm": 4.8090131934626115,
"learning_rate": 1e-06,
"loss": 0.4777,
"step": 2040
},
{
"epoch": 0.7804971319311663,
"grad_norm": 2.7526655283536225,
"learning_rate": 1e-06,
"loss": 0.4588,
"step": 2041
},
{
"epoch": 0.7808795411089866,
"grad_norm": 3.68605530450721,
"learning_rate": 1e-06,
"loss": 0.4125,
"step": 2042
},
{
"epoch": 0.7812619502868069,
"grad_norm": 2.9814896141381855,
"learning_rate": 1e-06,
"loss": 0.24,
"step": 2043
},
{
"epoch": 0.7816443594646272,
"grad_norm": 2.883457349487923,
"learning_rate": 1e-06,
"loss": 0.2844,
"step": 2044
},
{
"epoch": 0.7820267686424475,
"grad_norm": 2.7928683603024536,
"learning_rate": 1e-06,
"loss": 0.2646,
"step": 2045
},
{
"epoch": 0.7824091778202676,
"grad_norm": 3.7991790687831086,
"learning_rate": 1e-06,
"loss": 0.4261,
"step": 2046
},
{
"epoch": 0.7827915869980879,
"grad_norm": 2.2591240581273455,
"learning_rate": 1e-06,
"loss": 0.1932,
"step": 2047
},
{
"epoch": 0.7831739961759082,
"grad_norm": 4.054189780085439,
"learning_rate": 1e-06,
"loss": 0.1903,
"step": 2048
},
{
"epoch": 0.7835564053537285,
"grad_norm": 2.4492055110786533,
"learning_rate": 1e-06,
"loss": 0.0825,
"step": 2049
},
{
"epoch": 0.7839388145315488,
"grad_norm": 4.351313969142728,
"learning_rate": 1e-06,
"loss": 0.0969,
"step": 2050
},
{
"epoch": 0.7843212237093691,
"grad_norm": 5.379664326799858,
"learning_rate": 1e-06,
"loss": 0.6871,
"step": 2051
},
{
"epoch": 0.7847036328871893,
"grad_norm": 2.9534803711545825,
"learning_rate": 1e-06,
"loss": 0.4627,
"step": 2052
},
{
"epoch": 0.7850860420650095,
"grad_norm": 2.7381981852036636,
"learning_rate": 1e-06,
"loss": 0.3545,
"step": 2053
},
{
"epoch": 0.7854684512428298,
"grad_norm": 2.803899202632896,
"learning_rate": 1e-06,
"loss": 0.3812,
"step": 2054
},
{
"epoch": 0.7858508604206501,
"grad_norm": 4.112794331044607,
"learning_rate": 1e-06,
"loss": 0.4038,
"step": 2055
},
{
"epoch": 0.7862332695984704,
"grad_norm": 2.702379950892697,
"learning_rate": 1e-06,
"loss": 0.1801,
"step": 2056
},
{
"epoch": 0.7866156787762907,
"grad_norm": 2.579214154999277,
"learning_rate": 1e-06,
"loss": 0.249,
"step": 2057
},
{
"epoch": 0.7869980879541109,
"grad_norm": 2.8267955711350563,
"learning_rate": 1e-06,
"loss": 0.1629,
"step": 2058
},
{
"epoch": 0.7873804971319311,
"grad_norm": 3.2930302913974288,
"learning_rate": 1e-06,
"loss": 0.0863,
"step": 2059
},
{
"epoch": 0.7877629063097514,
"grad_norm": 3.6787594153296386,
"learning_rate": 1e-06,
"loss": 0.145,
"step": 2060
},
{
"epoch": 0.7881453154875717,
"grad_norm": 3.5522703567148186,
"learning_rate": 1e-06,
"loss": 0.1292,
"step": 2061
},
{
"epoch": 0.788527724665392,
"grad_norm": 4.591228182164673,
"learning_rate": 1e-06,
"loss": 0.1309,
"step": 2062
},
{
"epoch": 0.7889101338432123,
"grad_norm": 3.541918682032985,
"learning_rate": 1e-06,
"loss": 0.2255,
"step": 2063
},
{
"epoch": 0.7892925430210325,
"grad_norm": 4.971798519407049,
"learning_rate": 1e-06,
"loss": 0.471,
"step": 2064
},
{
"epoch": 0.7896749521988528,
"grad_norm": 4.940052193740455,
"learning_rate": 1e-06,
"loss": 0.4142,
"step": 2065
},
{
"epoch": 0.790057361376673,
"grad_norm": 4.393978914942134,
"learning_rate": 1e-06,
"loss": 0.3067,
"step": 2066
},
{
"epoch": 0.7904397705544933,
"grad_norm": 4.102075860730455,
"learning_rate": 1e-06,
"loss": 0.3246,
"step": 2067
},
{
"epoch": 0.7908221797323136,
"grad_norm": 2.186783046397264,
"learning_rate": 1e-06,
"loss": 0.3369,
"step": 2068
},
{
"epoch": 0.7912045889101338,
"grad_norm": 2.822973459571022,
"learning_rate": 1e-06,
"loss": 0.2647,
"step": 2069
},
{
"epoch": 0.7915869980879541,
"grad_norm": 3.3141656682396077,
"learning_rate": 1e-06,
"loss": 0.3238,
"step": 2070
},
{
"epoch": 0.7919694072657744,
"grad_norm": 2.6746682184119144,
"learning_rate": 1e-06,
"loss": 0.194,
"step": 2071
},
{
"epoch": 0.7923518164435946,
"grad_norm": 8.472692268826158,
"learning_rate": 1e-06,
"loss": 0.3371,
"step": 2072
},
{
"epoch": 0.7927342256214149,
"grad_norm": 3.716701039267943,
"learning_rate": 1e-06,
"loss": 0.1934,
"step": 2073
},
{
"epoch": 0.7931166347992352,
"grad_norm": 4.896144689304383,
"learning_rate": 1e-06,
"loss": 0.1346,
"step": 2074
},
{
"epoch": 0.7934990439770554,
"grad_norm": 4.428809884678829,
"learning_rate": 1e-06,
"loss": 0.2384,
"step": 2075
},
{
"epoch": 0.7938814531548757,
"grad_norm": 6.608011307953772,
"learning_rate": 1e-06,
"loss": 0.4356,
"step": 2076
},
{
"epoch": 0.794263862332696,
"grad_norm": 3.8141769723306393,
"learning_rate": 1e-06,
"loss": 0.4045,
"step": 2077
},
{
"epoch": 0.7946462715105163,
"grad_norm": 2.7864923683060625,
"learning_rate": 1e-06,
"loss": 0.323,
"step": 2078
},
{
"epoch": 0.7950286806883365,
"grad_norm": 2.414149705614162,
"learning_rate": 1e-06,
"loss": 0.379,
"step": 2079
},
{
"epoch": 0.7954110898661568,
"grad_norm": 3.170569841930405,
"learning_rate": 1e-06,
"loss": 0.2751,
"step": 2080
},
{
"epoch": 0.795793499043977,
"grad_norm": 3.8556209947698514,
"learning_rate": 1e-06,
"loss": 0.4716,
"step": 2081
},
{
"epoch": 0.7961759082217973,
"grad_norm": 2.704039490642762,
"learning_rate": 1e-06,
"loss": 0.4642,
"step": 2082
},
{
"epoch": 0.7965583173996176,
"grad_norm": 2.478727339171121,
"learning_rate": 1e-06,
"loss": 0.2617,
"step": 2083
},
{
"epoch": 0.7969407265774379,
"grad_norm": 3.615397923096984,
"learning_rate": 1e-06,
"loss": 0.2749,
"step": 2084
},
{
"epoch": 0.7973231357552581,
"grad_norm": 3.7581430396873516,
"learning_rate": 1e-06,
"loss": 0.2536,
"step": 2085
},
{
"epoch": 0.7977055449330784,
"grad_norm": 3.1040276788578613,
"learning_rate": 1e-06,
"loss": 0.1833,
"step": 2086
},
{
"epoch": 0.7980879541108986,
"grad_norm": 3.7896379486113414,
"learning_rate": 1e-06,
"loss": 0.0956,
"step": 2087
},
{
"epoch": 0.7984703632887189,
"grad_norm": 6.420381172724574,
"learning_rate": 1e-06,
"loss": 0.1931,
"step": 2088
},
{
"epoch": 0.7988527724665392,
"grad_norm": 5.77144670038465,
"learning_rate": 1e-06,
"loss": 0.6335,
"step": 2089
},
{
"epoch": 0.7992351816443595,
"grad_norm": 4.792139157205931,
"learning_rate": 1e-06,
"loss": 0.3828,
"step": 2090
},
{
"epoch": 0.7996175908221798,
"grad_norm": 2.162194061688356,
"learning_rate": 1e-06,
"loss": 0.3287,
"step": 2091
},
{
"epoch": 0.8,
"grad_norm": 2.8522792803652104,
"learning_rate": 1e-06,
"loss": 0.3452,
"step": 2092
},
{
"epoch": 0.8003824091778202,
"grad_norm": 3.708971775831023,
"learning_rate": 1e-06,
"loss": 0.3769,
"step": 2093
},
{
"epoch": 0.8007648183556405,
"grad_norm": 3.4264057421019594,
"learning_rate": 1e-06,
"loss": 0.5301,
"step": 2094
},
{
"epoch": 0.8011472275334608,
"grad_norm": 2.1893933413099647,
"learning_rate": 1e-06,
"loss": 0.1319,
"step": 2095
},
{
"epoch": 0.8015296367112811,
"grad_norm": 1.9048559375667495,
"learning_rate": 1e-06,
"loss": 0.1312,
"step": 2096
},
{
"epoch": 0.8019120458891014,
"grad_norm": 3.8689251935544466,
"learning_rate": 1e-06,
"loss": 0.3397,
"step": 2097
},
{
"epoch": 0.8022944550669217,
"grad_norm": 5.43743352739766,
"learning_rate": 1e-06,
"loss": 0.3189,
"step": 2098
},
{
"epoch": 0.8026768642447418,
"grad_norm": 3.5664840639614823,
"learning_rate": 1e-06,
"loss": 0.1821,
"step": 2099
},
{
"epoch": 0.8030592734225621,
"grad_norm": 4.467345116934534,
"learning_rate": 1e-06,
"loss": 0.1286,
"step": 2100
},
{
"epoch": 0.8034416826003824,
"grad_norm": 5.400395393024568,
"learning_rate": 1e-06,
"loss": 0.4901,
"step": 2101
},
{
"epoch": 0.8038240917782027,
"grad_norm": 2.9126148053901044,
"learning_rate": 1e-06,
"loss": 0.3685,
"step": 2102
},
{
"epoch": 0.804206500956023,
"grad_norm": 3.0631492082331566,
"learning_rate": 1e-06,
"loss": 0.371,
"step": 2103
},
{
"epoch": 0.8045889101338433,
"grad_norm": 3.255761248615045,
"learning_rate": 1e-06,
"loss": 0.4381,
"step": 2104
},
{
"epoch": 0.8049713193116634,
"grad_norm": 2.3928423801771963,
"learning_rate": 1e-06,
"loss": 0.2927,
"step": 2105
},
{
"epoch": 0.8053537284894837,
"grad_norm": 5.051747523942017,
"learning_rate": 1e-06,
"loss": 0.5989,
"step": 2106
},
{
"epoch": 0.805736137667304,
"grad_norm": 3.504058664466598,
"learning_rate": 1e-06,
"loss": 0.198,
"step": 2107
},
{
"epoch": 0.8061185468451243,
"grad_norm": 3.1708870086604937,
"learning_rate": 1e-06,
"loss": 0.1881,
"step": 2108
},
{
"epoch": 0.8065009560229446,
"grad_norm": 2.917293118003835,
"learning_rate": 1e-06,
"loss": 0.2526,
"step": 2109
},
{
"epoch": 0.8068833652007649,
"grad_norm": 3.846004166258144,
"learning_rate": 1e-06,
"loss": 0.2796,
"step": 2110
},
{
"epoch": 0.807265774378585,
"grad_norm": 3.0444224341340176,
"learning_rate": 1e-06,
"loss": 0.1437,
"step": 2111
},
{
"epoch": 0.8076481835564053,
"grad_norm": 3.5297505393779964,
"learning_rate": 1e-06,
"loss": 0.1154,
"step": 2112
},
{
"epoch": 0.8080305927342256,
"grad_norm": 4.927851847527151,
"learning_rate": 1e-06,
"loss": 0.1744,
"step": 2113
},
{
"epoch": 0.8084130019120459,
"grad_norm": 4.281335314715068,
"learning_rate": 1e-06,
"loss": 0.3368,
"step": 2114
},
{
"epoch": 0.8087954110898662,
"grad_norm": 5.608802702460388,
"learning_rate": 1e-06,
"loss": 0.5034,
"step": 2115
},
{
"epoch": 0.8091778202676865,
"grad_norm": 2.787287898183835,
"learning_rate": 1e-06,
"loss": 0.406,
"step": 2116
},
{
"epoch": 0.8095602294455067,
"grad_norm": 1.6220800134229783,
"learning_rate": 1e-06,
"loss": 0.2047,
"step": 2117
},
{
"epoch": 0.8099426386233269,
"grad_norm": 2.822382370801178,
"learning_rate": 1e-06,
"loss": 0.2202,
"step": 2118
},
{
"epoch": 0.8103250478011472,
"grad_norm": 4.004621696750148,
"learning_rate": 1e-06,
"loss": 0.3022,
"step": 2119
},
{
"epoch": 0.8107074569789675,
"grad_norm": 4.1272121624366065,
"learning_rate": 1e-06,
"loss": 0.3423,
"step": 2120
},
{
"epoch": 0.8110898661567878,
"grad_norm": 2.2668283358120718,
"learning_rate": 1e-06,
"loss": 0.1929,
"step": 2121
},
{
"epoch": 0.8114722753346081,
"grad_norm": 4.1438659651742,
"learning_rate": 1e-06,
"loss": 0.2024,
"step": 2122
},
{
"epoch": 0.8118546845124283,
"grad_norm": 3.14419428755532,
"learning_rate": 1e-06,
"loss": 0.1916,
"step": 2123
},
{
"epoch": 0.8122370936902485,
"grad_norm": 3.3102456825100854,
"learning_rate": 1e-06,
"loss": 0.1257,
"step": 2124
},
{
"epoch": 0.8126195028680688,
"grad_norm": 3.8880439990972344,
"learning_rate": 1e-06,
"loss": 0.1195,
"step": 2125
},
{
"epoch": 0.8130019120458891,
"grad_norm": 7.244368635195714,
"learning_rate": 1e-06,
"loss": 0.6411,
"step": 2126
},
{
"epoch": 0.8133843212237094,
"grad_norm": 3.16347289066628,
"learning_rate": 1e-06,
"loss": 0.3795,
"step": 2127
},
{
"epoch": 0.8137667304015297,
"grad_norm": 3.2375683339098558,
"learning_rate": 1e-06,
"loss": 0.3666,
"step": 2128
},
{
"epoch": 0.8141491395793499,
"grad_norm": 2.5387238731523696,
"learning_rate": 1e-06,
"loss": 0.428,
"step": 2129
},
{
"epoch": 0.8145315487571702,
"grad_norm": 2.6485824320249045,
"learning_rate": 1e-06,
"loss": 0.3104,
"step": 2130
},
{
"epoch": 0.8149139579349904,
"grad_norm": 3.061018916513091,
"learning_rate": 1e-06,
"loss": 0.3034,
"step": 2131
},
{
"epoch": 0.8152963671128107,
"grad_norm": 2.5765817358109753,
"learning_rate": 1e-06,
"loss": 0.335,
"step": 2132
},
{
"epoch": 0.815678776290631,
"grad_norm": 3.6465646646054988,
"learning_rate": 1e-06,
"loss": 0.2214,
"step": 2133
},
{
"epoch": 0.8160611854684512,
"grad_norm": 3.659063446829306,
"learning_rate": 1e-06,
"loss": 0.2679,
"step": 2134
},
{
"epoch": 0.8164435946462715,
"grad_norm": 4.176751076793435,
"learning_rate": 1e-06,
"loss": 0.2274,
"step": 2135
},
{
"epoch": 0.8168260038240918,
"grad_norm": 2.9000141143455354,
"learning_rate": 1e-06,
"loss": 0.1416,
"step": 2136
},
{
"epoch": 0.817208413001912,
"grad_norm": 3.5229236033480564,
"learning_rate": 1e-06,
"loss": 0.1278,
"step": 2137
},
{
"epoch": 0.8175908221797323,
"grad_norm": 7.9440475252080205,
"learning_rate": 1e-06,
"loss": 0.242,
"step": 2138
},
{
"epoch": 0.8179732313575526,
"grad_norm": 5.1135330313058365,
"learning_rate": 1e-06,
"loss": 0.6503,
"step": 2139
},
{
"epoch": 0.8183556405353728,
"grad_norm": 4.578243527489235,
"learning_rate": 1e-06,
"loss": 0.5686,
"step": 2140
},
{
"epoch": 0.8187380497131931,
"grad_norm": 4.147331133365789,
"learning_rate": 1e-06,
"loss": 0.5776,
"step": 2141
},
{
"epoch": 0.8191204588910134,
"grad_norm": 1.9505210913813733,
"learning_rate": 1e-06,
"loss": 0.4936,
"step": 2142
},
{
"epoch": 0.8195028680688337,
"grad_norm": 3.321460911231512,
"learning_rate": 1e-06,
"loss": 0.2061,
"step": 2143
},
{
"epoch": 0.8198852772466539,
"grad_norm": 3.6699691696327217,
"learning_rate": 1e-06,
"loss": 0.2966,
"step": 2144
},
{
"epoch": 0.8202676864244742,
"grad_norm": 3.001923103653967,
"learning_rate": 1e-06,
"loss": 0.2617,
"step": 2145
},
{
"epoch": 0.8206500956022944,
"grad_norm": 2.4060394393149123,
"learning_rate": 1e-06,
"loss": 0.2851,
"step": 2146
},
{
"epoch": 0.8210325047801147,
"grad_norm": 3.218550592560278,
"learning_rate": 1e-06,
"loss": 0.2154,
"step": 2147
},
{
"epoch": 0.821414913957935,
"grad_norm": 4.372564237197992,
"learning_rate": 1e-06,
"loss": 0.2861,
"step": 2148
},
{
"epoch": 0.8217973231357553,
"grad_norm": 3.943080882818625,
"learning_rate": 1e-06,
"loss": 0.2615,
"step": 2149
},
{
"epoch": 0.8221797323135756,
"grad_norm": 4.047737412632037,
"learning_rate": 1e-06,
"loss": 0.2142,
"step": 2150
},
{
"epoch": 0.8225621414913958,
"grad_norm": 7.329343259359967,
"learning_rate": 1e-06,
"loss": 0.6134,
"step": 2151
},
{
"epoch": 0.822944550669216,
"grad_norm": 4.629531908204858,
"learning_rate": 1e-06,
"loss": 0.6396,
"step": 2152
},
{
"epoch": 0.8233269598470363,
"grad_norm": 4.350650054767431,
"learning_rate": 1e-06,
"loss": 0.4429,
"step": 2153
},
{
"epoch": 0.8237093690248566,
"grad_norm": 3.1065780865506847,
"learning_rate": 1e-06,
"loss": 0.248,
"step": 2154
},
{
"epoch": 0.8240917782026769,
"grad_norm": 3.943895746551621,
"learning_rate": 1e-06,
"loss": 0.3949,
"step": 2155
},
{
"epoch": 0.8244741873804972,
"grad_norm": 3.179876148523129,
"learning_rate": 1e-06,
"loss": 0.5421,
"step": 2156
},
{
"epoch": 0.8248565965583174,
"grad_norm": 3.589649946524929,
"learning_rate": 1e-06,
"loss": 0.2262,
"step": 2157
},
{
"epoch": 0.8252390057361376,
"grad_norm": 3.8537420640357034,
"learning_rate": 1e-06,
"loss": 0.3552,
"step": 2158
},
{
"epoch": 0.8256214149139579,
"grad_norm": 2.8540038644808594,
"learning_rate": 1e-06,
"loss": 0.2076,
"step": 2159
},
{
"epoch": 0.8260038240917782,
"grad_norm": 3.918334578742821,
"learning_rate": 1e-06,
"loss": 0.3502,
"step": 2160
},
{
"epoch": 0.8263862332695985,
"grad_norm": 4.105354459933907,
"learning_rate": 1e-06,
"loss": 0.1463,
"step": 2161
},
{
"epoch": 0.8267686424474188,
"grad_norm": 4.794980547948051,
"learning_rate": 1e-06,
"loss": 0.1324,
"step": 2162
},
{
"epoch": 0.827151051625239,
"grad_norm": 5.959562410509337,
"learning_rate": 1e-06,
"loss": 0.2033,
"step": 2163
},
{
"epoch": 0.8275334608030592,
"grad_norm": 4.1375193339008485,
"learning_rate": 1e-06,
"loss": 0.5293,
"step": 2164
},
{
"epoch": 0.8279158699808795,
"grad_norm": 3.7650827357711742,
"learning_rate": 1e-06,
"loss": 0.3524,
"step": 2165
},
{
"epoch": 0.8282982791586998,
"grad_norm": 2.258227247535923,
"learning_rate": 1e-06,
"loss": 0.1928,
"step": 2166
},
{
"epoch": 0.8286806883365201,
"grad_norm": 2.73280594630298,
"learning_rate": 1e-06,
"loss": 0.3709,
"step": 2167
},
{
"epoch": 0.8290630975143404,
"grad_norm": 3.8576784922553324,
"learning_rate": 1e-06,
"loss": 0.4781,
"step": 2168
},
{
"epoch": 0.8294455066921607,
"grad_norm": 2.629339355343209,
"learning_rate": 1e-06,
"loss": 0.384,
"step": 2169
},
{
"epoch": 0.8298279158699808,
"grad_norm": 3.0564081755663697,
"learning_rate": 1e-06,
"loss": 0.3269,
"step": 2170
},
{
"epoch": 0.8302103250478011,
"grad_norm": 2.679405798159991,
"learning_rate": 1e-06,
"loss": 0.2193,
"step": 2171
},
{
"epoch": 0.8305927342256214,
"grad_norm": 3.185608096677568,
"learning_rate": 1e-06,
"loss": 0.2305,
"step": 2172
},
{
"epoch": 0.8309751434034417,
"grad_norm": 4.0877777098595525,
"learning_rate": 1e-06,
"loss": 0.2141,
"step": 2173
},
{
"epoch": 0.831357552581262,
"grad_norm": 3.0768842181172853,
"learning_rate": 1e-06,
"loss": 0.1719,
"step": 2174
},
{
"epoch": 0.8317399617590823,
"grad_norm": 3.5653037030946315,
"learning_rate": 1e-06,
"loss": 0.1642,
"step": 2175
},
{
"epoch": 0.8321223709369024,
"grad_norm": 3.9048649888382423,
"learning_rate": 1e-06,
"loss": 0.3858,
"step": 2176
},
{
"epoch": 0.8325047801147227,
"grad_norm": 3.1362740663446607,
"learning_rate": 1e-06,
"loss": 0.5031,
"step": 2177
},
{
"epoch": 0.832887189292543,
"grad_norm": 3.297428898135157,
"learning_rate": 1e-06,
"loss": 0.6066,
"step": 2178
},
{
"epoch": 0.8332695984703633,
"grad_norm": 2.465040776377811,
"learning_rate": 1e-06,
"loss": 0.3532,
"step": 2179
},
{
"epoch": 0.8336520076481836,
"grad_norm": 2.667095348551791,
"learning_rate": 1e-06,
"loss": 0.3922,
"step": 2180
},
{
"epoch": 0.8340344168260039,
"grad_norm": 2.9862716000822846,
"learning_rate": 1e-06,
"loss": 0.2624,
"step": 2181
},
{
"epoch": 0.834416826003824,
"grad_norm": 2.5867503752715075,
"learning_rate": 1e-06,
"loss": 0.2768,
"step": 2182
},
{
"epoch": 0.8347992351816443,
"grad_norm": 3.99836637993455,
"learning_rate": 1e-06,
"loss": 0.3965,
"step": 2183
},
{
"epoch": 0.8351816443594646,
"grad_norm": 3.9710708200234355,
"learning_rate": 1e-06,
"loss": 0.2693,
"step": 2184
},
{
"epoch": 0.8355640535372849,
"grad_norm": 3.4491208711692822,
"learning_rate": 1e-06,
"loss": 0.2509,
"step": 2185
},
{
"epoch": 0.8359464627151052,
"grad_norm": 2.6952109497400447,
"learning_rate": 1e-06,
"loss": 0.1052,
"step": 2186
},
{
"epoch": 0.8363288718929255,
"grad_norm": 3.271212237981573,
"learning_rate": 1e-06,
"loss": 0.1229,
"step": 2187
},
{
"epoch": 0.8367112810707457,
"grad_norm": 5.315335762277588,
"learning_rate": 1e-06,
"loss": 0.3858,
"step": 2188
},
{
"epoch": 0.837093690248566,
"grad_norm": 4.039271690514878,
"learning_rate": 1e-06,
"loss": 0.3017,
"step": 2189
},
{
"epoch": 0.8374760994263862,
"grad_norm": 4.276716742000643,
"learning_rate": 1e-06,
"loss": 0.4611,
"step": 2190
},
{
"epoch": 0.8378585086042065,
"grad_norm": 2.3208846597343618,
"learning_rate": 1e-06,
"loss": 0.2205,
"step": 2191
},
{
"epoch": 0.8382409177820268,
"grad_norm": 2.910222118067495,
"learning_rate": 1e-06,
"loss": 0.3472,
"step": 2192
},
{
"epoch": 0.838623326959847,
"grad_norm": 3.7821854624348363,
"learning_rate": 1e-06,
"loss": 0.4182,
"step": 2193
},
{
"epoch": 0.8390057361376673,
"grad_norm": 3.645645412869907,
"learning_rate": 1e-06,
"loss": 0.4144,
"step": 2194
},
{
"epoch": 0.8393881453154876,
"grad_norm": 5.041450159557636,
"learning_rate": 1e-06,
"loss": 0.279,
"step": 2195
},
{
"epoch": 0.8397705544933078,
"grad_norm": 2.872907457528259,
"learning_rate": 1e-06,
"loss": 0.197,
"step": 2196
},
{
"epoch": 0.8401529636711281,
"grad_norm": 2.217400046141906,
"learning_rate": 1e-06,
"loss": 0.1678,
"step": 2197
},
{
"epoch": 0.8405353728489484,
"grad_norm": 5.82892511106803,
"learning_rate": 1e-06,
"loss": 0.256,
"step": 2198
},
{
"epoch": 0.8409177820267686,
"grad_norm": 7.5717419356407945,
"learning_rate": 1e-06,
"loss": 0.1019,
"step": 2199
},
{
"epoch": 0.8413001912045889,
"grad_norm": 3.061982091590941,
"learning_rate": 1e-06,
"loss": 0.0776,
"step": 2200
},
{
"epoch": 0.8416826003824092,
"grad_norm": 4.486253723670178,
"learning_rate": 1e-06,
"loss": 0.3442,
"step": 2201
},
{
"epoch": 0.8420650095602294,
"grad_norm": 3.257880267346353,
"learning_rate": 1e-06,
"loss": 0.4429,
"step": 2202
},
{
"epoch": 0.8424474187380497,
"grad_norm": 2.2060811656293344,
"learning_rate": 1e-06,
"loss": 0.3688,
"step": 2203
},
{
"epoch": 0.84282982791587,
"grad_norm": 2.597016265735945,
"learning_rate": 1e-06,
"loss": 0.2879,
"step": 2204
},
{
"epoch": 0.8432122370936902,
"grad_norm": 2.395171983632231,
"learning_rate": 1e-06,
"loss": 0.2489,
"step": 2205
},
{
"epoch": 0.8435946462715105,
"grad_norm": 3.1052566935755257,
"learning_rate": 1e-06,
"loss": 0.3973,
"step": 2206
},
{
"epoch": 0.8439770554493308,
"grad_norm": 3.2083941036925894,
"learning_rate": 1e-06,
"loss": 0.265,
"step": 2207
},
{
"epoch": 0.8443594646271511,
"grad_norm": 3.424640809318776,
"learning_rate": 1e-06,
"loss": 0.1639,
"step": 2208
},
{
"epoch": 0.8447418738049713,
"grad_norm": 4.030188132312924,
"learning_rate": 1e-06,
"loss": 0.2207,
"step": 2209
},
{
"epoch": 0.8451242829827916,
"grad_norm": 2.6657795225124086,
"learning_rate": 1e-06,
"loss": 0.101,
"step": 2210
},
{
"epoch": 0.8455066921606118,
"grad_norm": 3.655967766717767,
"learning_rate": 1e-06,
"loss": 0.1517,
"step": 2211
},
{
"epoch": 0.8458891013384321,
"grad_norm": 4.780504156044833,
"learning_rate": 1e-06,
"loss": 0.2003,
"step": 2212
},
{
"epoch": 0.8462715105162524,
"grad_norm": 7.961990661140336,
"learning_rate": 1e-06,
"loss": 0.1447,
"step": 2213
},
{
"epoch": 0.8466539196940727,
"grad_norm": 6.391086104716505,
"learning_rate": 1e-06,
"loss": 0.4777,
"step": 2214
},
{
"epoch": 0.847036328871893,
"grad_norm": 3.4270184326210464,
"learning_rate": 1e-06,
"loss": 0.3738,
"step": 2215
},
{
"epoch": 0.8474187380497132,
"grad_norm": 2.598951524459288,
"learning_rate": 1e-06,
"loss": 0.3055,
"step": 2216
},
{
"epoch": 0.8478011472275334,
"grad_norm": 2.271296276068281,
"learning_rate": 1e-06,
"loss": 0.3695,
"step": 2217
},
{
"epoch": 0.8481835564053537,
"grad_norm": 4.049813753719958,
"learning_rate": 1e-06,
"loss": 0.5607,
"step": 2218
},
{
"epoch": 0.848565965583174,
"grad_norm": 2.4192822760293224,
"learning_rate": 1e-06,
"loss": 0.2376,
"step": 2219
},
{
"epoch": 0.8489483747609943,
"grad_norm": 2.7489896131911116,
"learning_rate": 1e-06,
"loss": 0.4236,
"step": 2220
},
{
"epoch": 0.8493307839388146,
"grad_norm": 2.5501943850758444,
"learning_rate": 1e-06,
"loss": 0.2919,
"step": 2221
},
{
"epoch": 0.8497131931166348,
"grad_norm": 2.2434257251745597,
"learning_rate": 1e-06,
"loss": 0.1412,
"step": 2222
},
{
"epoch": 0.850095602294455,
"grad_norm": 4.330815439594046,
"learning_rate": 1e-06,
"loss": 0.2933,
"step": 2223
},
{
"epoch": 0.8504780114722753,
"grad_norm": 5.112670022613608,
"learning_rate": 1e-06,
"loss": 0.1171,
"step": 2224
},
{
"epoch": 0.8508604206500956,
"grad_norm": 3.269080419233297,
"learning_rate": 1e-06,
"loss": 0.1149,
"step": 2225
},
{
"epoch": 0.8512428298279159,
"grad_norm": 4.993048604088785,
"learning_rate": 1e-06,
"loss": 0.5126,
"step": 2226
},
{
"epoch": 0.8516252390057362,
"grad_norm": 3.5827526834504617,
"learning_rate": 1e-06,
"loss": 0.6183,
"step": 2227
},
{
"epoch": 0.8520076481835565,
"grad_norm": 4.147780659143189,
"learning_rate": 1e-06,
"loss": 0.3283,
"step": 2228
},
{
"epoch": 0.8523900573613766,
"grad_norm": 2.6242208686949526,
"learning_rate": 1e-06,
"loss": 0.4896,
"step": 2229
},
{
"epoch": 0.8527724665391969,
"grad_norm": 2.682157306947873,
"learning_rate": 1e-06,
"loss": 0.3766,
"step": 2230
},
{
"epoch": 0.8531548757170172,
"grad_norm": 3.256507374262415,
"learning_rate": 1e-06,
"loss": 0.3188,
"step": 2231
},
{
"epoch": 0.8535372848948375,
"grad_norm": 3.0808107693662503,
"learning_rate": 1e-06,
"loss": 0.2812,
"step": 2232
},
{
"epoch": 0.8539196940726578,
"grad_norm": 2.9379973294925694,
"learning_rate": 1e-06,
"loss": 0.3041,
"step": 2233
},
{
"epoch": 0.8543021032504781,
"grad_norm": 2.3469328567139156,
"learning_rate": 1e-06,
"loss": 0.1998,
"step": 2234
},
{
"epoch": 0.8546845124282982,
"grad_norm": 3.5654014684500237,
"learning_rate": 1e-06,
"loss": 0.1945,
"step": 2235
},
{
"epoch": 0.8550669216061185,
"grad_norm": 2.794885620479178,
"learning_rate": 1e-06,
"loss": 0.1304,
"step": 2236
},
{
"epoch": 0.8554493307839388,
"grad_norm": 3.892644151432938,
"learning_rate": 1e-06,
"loss": 0.1376,
"step": 2237
},
{
"epoch": 0.8558317399617591,
"grad_norm": 4.52290069610058,
"learning_rate": 1e-06,
"loss": 0.322,
"step": 2238
},
{
"epoch": 0.8562141491395794,
"grad_norm": 4.4033401295919665,
"learning_rate": 1e-06,
"loss": 0.4721,
"step": 2239
},
{
"epoch": 0.8565965583173997,
"grad_norm": 2.8414582460400957,
"learning_rate": 1e-06,
"loss": 0.2592,
"step": 2240
},
{
"epoch": 0.8569789674952198,
"grad_norm": 3.4026742067383045,
"learning_rate": 1e-06,
"loss": 0.283,
"step": 2241
},
{
"epoch": 0.8573613766730401,
"grad_norm": 2.563220783370359,
"learning_rate": 1e-06,
"loss": 0.2566,
"step": 2242
},
{
"epoch": 0.8577437858508604,
"grad_norm": 1.9607244733019165,
"learning_rate": 1e-06,
"loss": 0.4699,
"step": 2243
},
{
"epoch": 0.8581261950286807,
"grad_norm": 3.702986573798893,
"learning_rate": 1e-06,
"loss": 0.2381,
"step": 2244
},
{
"epoch": 0.858508604206501,
"grad_norm": 2.3633086904398657,
"learning_rate": 1e-06,
"loss": 0.3277,
"step": 2245
},
{
"epoch": 0.8588910133843213,
"grad_norm": 2.719397895171275,
"learning_rate": 1e-06,
"loss": 0.1863,
"step": 2246
},
{
"epoch": 0.8592734225621415,
"grad_norm": 4.029146577192611,
"learning_rate": 1e-06,
"loss": 0.2504,
"step": 2247
},
{
"epoch": 0.8596558317399617,
"grad_norm": 2.9593021436220557,
"learning_rate": 1e-06,
"loss": 0.2487,
"step": 2248
},
{
"epoch": 0.860038240917782,
"grad_norm": 4.2307703645077765,
"learning_rate": 1e-06,
"loss": 0.2118,
"step": 2249
},
{
"epoch": 0.8604206500956023,
"grad_norm": 5.035898276758321,
"learning_rate": 1e-06,
"loss": 0.2637,
"step": 2250
},
{
"epoch": 0.8608030592734226,
"grad_norm": 8.803036200658138,
"learning_rate": 1e-06,
"loss": 0.504,
"step": 2251
},
{
"epoch": 0.8611854684512428,
"grad_norm": 4.752974131219422,
"learning_rate": 1e-06,
"loss": 0.6026,
"step": 2252
},
{
"epoch": 0.8615678776290631,
"grad_norm": 3.3525969502812525,
"learning_rate": 1e-06,
"loss": 0.4438,
"step": 2253
},
{
"epoch": 0.8619502868068833,
"grad_norm": 3.616882179507573,
"learning_rate": 1e-06,
"loss": 0.2969,
"step": 2254
},
{
"epoch": 0.8623326959847036,
"grad_norm": 2.2632922561056605,
"learning_rate": 1e-06,
"loss": 0.4263,
"step": 2255
},
{
"epoch": 0.8627151051625239,
"grad_norm": 2.3040020456768597,
"learning_rate": 1e-06,
"loss": 0.1578,
"step": 2256
},
{
"epoch": 0.8630975143403442,
"grad_norm": 2.0374035865111786,
"learning_rate": 1e-06,
"loss": 0.2902,
"step": 2257
},
{
"epoch": 0.8634799235181644,
"grad_norm": 3.4553879447456906,
"learning_rate": 1e-06,
"loss": 0.1156,
"step": 2258
},
{
"epoch": 0.8638623326959847,
"grad_norm": 3.5985599339635344,
"learning_rate": 1e-06,
"loss": 0.2264,
"step": 2259
},
{
"epoch": 0.864244741873805,
"grad_norm": 3.138347198272023,
"learning_rate": 1e-06,
"loss": 0.1511,
"step": 2260
},
{
"epoch": 0.8646271510516252,
"grad_norm": 3.2196779857762614,
"learning_rate": 1e-06,
"loss": 0.1603,
"step": 2261
},
{
"epoch": 0.8650095602294455,
"grad_norm": 3.5260168137606565,
"learning_rate": 1e-06,
"loss": 0.1481,
"step": 2262
},
{
"epoch": 0.8653919694072658,
"grad_norm": 8.299671800857737,
"learning_rate": 1e-06,
"loss": 0.2358,
"step": 2263
},
{
"epoch": 0.865774378585086,
"grad_norm": 4.560272300018649,
"learning_rate": 1e-06,
"loss": 0.3339,
"step": 2264
},
{
"epoch": 0.8661567877629063,
"grad_norm": 3.5616443342604454,
"learning_rate": 1e-06,
"loss": 0.4314,
"step": 2265
},
{
"epoch": 0.8665391969407266,
"grad_norm": 2.1089186563301427,
"learning_rate": 1e-06,
"loss": 0.5122,
"step": 2266
},
{
"epoch": 0.8669216061185469,
"grad_norm": 2.3026888181986362,
"learning_rate": 1e-06,
"loss": 0.3382,
"step": 2267
},
{
"epoch": 0.8673040152963671,
"grad_norm": 3.6717550055193824,
"learning_rate": 1e-06,
"loss": 0.3318,
"step": 2268
},
{
"epoch": 0.8676864244741874,
"grad_norm": 4.314919718927125,
"learning_rate": 1e-06,
"loss": 0.3843,
"step": 2269
},
{
"epoch": 0.8680688336520076,
"grad_norm": 4.4882347164281615,
"learning_rate": 1e-06,
"loss": 0.3663,
"step": 2270
},
{
"epoch": 0.8684512428298279,
"grad_norm": 3.2509591081218883,
"learning_rate": 1e-06,
"loss": 0.3615,
"step": 2271
},
{
"epoch": 0.8688336520076482,
"grad_norm": 4.111744708704633,
"learning_rate": 1e-06,
"loss": 0.3047,
"step": 2272
},
{
"epoch": 0.8692160611854685,
"grad_norm": 2.742256098484252,
"learning_rate": 1e-06,
"loss": 0.0854,
"step": 2273
},
{
"epoch": 0.8695984703632887,
"grad_norm": 2.907756783978073,
"learning_rate": 1e-06,
"loss": 0.1346,
"step": 2274
},
{
"epoch": 0.869980879541109,
"grad_norm": 3.8665304839608385,
"learning_rate": 1e-06,
"loss": 0.1086,
"step": 2275
},
{
"epoch": 0.8703632887189292,
"grad_norm": 4.447655675349624,
"learning_rate": 1e-06,
"loss": 0.4916,
"step": 2276
},
{
"epoch": 0.8707456978967495,
"grad_norm": 3.674655721718259,
"learning_rate": 1e-06,
"loss": 0.5621,
"step": 2277
},
{
"epoch": 0.8711281070745698,
"grad_norm": 2.352703844049994,
"learning_rate": 1e-06,
"loss": 0.5073,
"step": 2278
},
{
"epoch": 0.8715105162523901,
"grad_norm": 2.4129684586374416,
"learning_rate": 1e-06,
"loss": 0.2702,
"step": 2279
},
{
"epoch": 0.8718929254302104,
"grad_norm": 2.4772123332611757,
"learning_rate": 1e-06,
"loss": 0.209,
"step": 2280
},
{
"epoch": 0.8722753346080306,
"grad_norm": 2.809172548121751,
"learning_rate": 1e-06,
"loss": 0.298,
"step": 2281
},
{
"epoch": 0.8726577437858508,
"grad_norm": 2.24072068780217,
"learning_rate": 1e-06,
"loss": 0.2951,
"step": 2282
},
{
"epoch": 0.8730401529636711,
"grad_norm": 2.7390198412117717,
"learning_rate": 1e-06,
"loss": 0.252,
"step": 2283
},
{
"epoch": 0.8734225621414914,
"grad_norm": 3.9408881732569134,
"learning_rate": 1e-06,
"loss": 0.4563,
"step": 2284
},
{
"epoch": 0.8738049713193117,
"grad_norm": 4.011042849052909,
"learning_rate": 1e-06,
"loss": 0.2125,
"step": 2285
},
{
"epoch": 0.874187380497132,
"grad_norm": 2.976018099924646,
"learning_rate": 1e-06,
"loss": 0.1403,
"step": 2286
},
{
"epoch": 0.8745697896749522,
"grad_norm": 4.713013047842722,
"learning_rate": 1e-06,
"loss": 0.1347,
"step": 2287
},
{
"epoch": 0.8749521988527724,
"grad_norm": 4.383286394416696,
"learning_rate": 1e-06,
"loss": 0.295,
"step": 2288
},
{
"epoch": 0.8753346080305927,
"grad_norm": 3.5377202804510763,
"learning_rate": 1e-06,
"loss": 0.5638,
"step": 2289
},
{
"epoch": 0.875717017208413,
"grad_norm": 2.5127476888002107,
"learning_rate": 1e-06,
"loss": 0.3803,
"step": 2290
},
{
"epoch": 0.8760994263862333,
"grad_norm": 3.7291951666796046,
"learning_rate": 1e-06,
"loss": 0.4362,
"step": 2291
},
{
"epoch": 0.8764818355640536,
"grad_norm": 2.9703860944513365,
"learning_rate": 1e-06,
"loss": 0.2383,
"step": 2292
},
{
"epoch": 0.8768642447418739,
"grad_norm": 3.187304920884007,
"learning_rate": 1e-06,
"loss": 0.3238,
"step": 2293
},
{
"epoch": 0.877246653919694,
"grad_norm": 2.690191916812979,
"learning_rate": 1e-06,
"loss": 0.3012,
"step": 2294
},
{
"epoch": 0.8776290630975143,
"grad_norm": 1.8738393370114195,
"learning_rate": 1e-06,
"loss": 0.2375,
"step": 2295
},
{
"epoch": 0.8780114722753346,
"grad_norm": 2.5574251491000695,
"learning_rate": 1e-06,
"loss": 0.2422,
"step": 2296
},
{
"epoch": 0.8783938814531549,
"grad_norm": 3.842897149970192,
"learning_rate": 1e-06,
"loss": 0.2012,
"step": 2297
},
{
"epoch": 0.8787762906309752,
"grad_norm": 4.2444895755793945,
"learning_rate": 1e-06,
"loss": 0.1554,
"step": 2298
},
{
"epoch": 0.8791586998087955,
"grad_norm": 3.3474364592466648,
"learning_rate": 1e-06,
"loss": 0.0742,
"step": 2299
},
{
"epoch": 0.8795411089866156,
"grad_norm": 4.249059011967546,
"learning_rate": 1e-06,
"loss": 0.146,
"step": 2300
},
{
"epoch": 0.8799235181644359,
"grad_norm": 8.497555549532043,
"learning_rate": 1e-06,
"loss": 0.4556,
"step": 2301
},
{
"epoch": 0.8803059273422562,
"grad_norm": 4.212337339073676,
"learning_rate": 1e-06,
"loss": 0.5643,
"step": 2302
},
{
"epoch": 0.8806883365200765,
"grad_norm": 3.865618023137525,
"learning_rate": 1e-06,
"loss": 0.508,
"step": 2303
},
{
"epoch": 0.8810707456978968,
"grad_norm": 2.7602690651256534,
"learning_rate": 1e-06,
"loss": 0.3783,
"step": 2304
},
{
"epoch": 0.8814531548757171,
"grad_norm": 1.8248915156689647,
"learning_rate": 1e-06,
"loss": 0.233,
"step": 2305
},
{
"epoch": 0.8818355640535372,
"grad_norm": 2.1531514557872695,
"learning_rate": 1e-06,
"loss": 0.3302,
"step": 2306
},
{
"epoch": 0.8822179732313575,
"grad_norm": 3.267504575532075,
"learning_rate": 1e-06,
"loss": 0.3309,
"step": 2307
},
{
"epoch": 0.8826003824091778,
"grad_norm": 3.3718184557425874,
"learning_rate": 1e-06,
"loss": 0.2072,
"step": 2308
},
{
"epoch": 0.8829827915869981,
"grad_norm": 2.665566565173866,
"learning_rate": 1e-06,
"loss": 0.0938,
"step": 2309
},
{
"epoch": 0.8833652007648184,
"grad_norm": 3.2153290281520714,
"learning_rate": 1e-06,
"loss": 0.2025,
"step": 2310
},
{
"epoch": 0.8837476099426386,
"grad_norm": 3.020818004350026,
"learning_rate": 1e-06,
"loss": 0.1081,
"step": 2311
},
{
"epoch": 0.8841300191204589,
"grad_norm": 5.98992423593946,
"learning_rate": 1e-06,
"loss": 0.1368,
"step": 2312
},
{
"epoch": 0.8845124282982791,
"grad_norm": 7.505415677752273,
"learning_rate": 1e-06,
"loss": 0.2759,
"step": 2313
},
{
"epoch": 0.8848948374760994,
"grad_norm": 4.147387700414422,
"learning_rate": 1e-06,
"loss": 0.382,
"step": 2314
},
{
"epoch": 0.8852772466539197,
"grad_norm": 2.1308112624179936,
"learning_rate": 1e-06,
"loss": 0.475,
"step": 2315
},
{
"epoch": 0.88565965583174,
"grad_norm": 3.5914450509956697,
"learning_rate": 1e-06,
"loss": 0.1266,
"step": 2316
},
{
"epoch": 0.8860420650095602,
"grad_norm": 3.7600621331945296,
"learning_rate": 1e-06,
"loss": 0.2785,
"step": 2317
},
{
"epoch": 0.8864244741873805,
"grad_norm": 2.826946455570886,
"learning_rate": 1e-06,
"loss": 0.3301,
"step": 2318
},
{
"epoch": 0.8868068833652007,
"grad_norm": 2.580528017911607,
"learning_rate": 1e-06,
"loss": 0.4009,
"step": 2319
},
{
"epoch": 0.887189292543021,
"grad_norm": 4.593738789447418,
"learning_rate": 1e-06,
"loss": 0.3341,
"step": 2320
},
{
"epoch": 0.8875717017208413,
"grad_norm": 3.9482181537961933,
"learning_rate": 1e-06,
"loss": 0.3531,
"step": 2321
},
{
"epoch": 0.8879541108986616,
"grad_norm": 3.688090778160383,
"learning_rate": 1e-06,
"loss": 0.1288,
"step": 2322
},
{
"epoch": 0.8883365200764818,
"grad_norm": 2.5997625315867006,
"learning_rate": 1e-06,
"loss": 0.0716,
"step": 2323
},
{
"epoch": 0.8887189292543021,
"grad_norm": 4.739358727579254,
"learning_rate": 1e-06,
"loss": 0.1075,
"step": 2324
},
{
"epoch": 0.8891013384321224,
"grad_norm": 3.958546535706509,
"learning_rate": 1e-06,
"loss": 0.1491,
"step": 2325
},
{
"epoch": 0.8894837476099426,
"grad_norm": 8.595346531387138,
"learning_rate": 1e-06,
"loss": 0.4131,
"step": 2326
},
{
"epoch": 0.8898661567877629,
"grad_norm": 3.994591871667418,
"learning_rate": 1e-06,
"loss": 0.3402,
"step": 2327
},
{
"epoch": 0.8902485659655832,
"grad_norm": 5.926634151911645,
"learning_rate": 1e-06,
"loss": 0.4476,
"step": 2328
},
{
"epoch": 0.8906309751434034,
"grad_norm": 3.0411437294890624,
"learning_rate": 1e-06,
"loss": 0.3332,
"step": 2329
},
{
"epoch": 0.8910133843212237,
"grad_norm": 5.351350432043598,
"learning_rate": 1e-06,
"loss": 0.3868,
"step": 2330
},
{
"epoch": 0.891395793499044,
"grad_norm": 4.472179670638214,
"learning_rate": 1e-06,
"loss": 0.3971,
"step": 2331
},
{
"epoch": 0.8917782026768643,
"grad_norm": 3.865687347131359,
"learning_rate": 1e-06,
"loss": 0.1836,
"step": 2332
},
{
"epoch": 0.8921606118546845,
"grad_norm": 2.8248239715557286,
"learning_rate": 1e-06,
"loss": 0.2198,
"step": 2333
},
{
"epoch": 0.8925430210325048,
"grad_norm": 2.7267736664679947,
"learning_rate": 1e-06,
"loss": 0.1597,
"step": 2334
},
{
"epoch": 0.892925430210325,
"grad_norm": 3.587110672403936,
"learning_rate": 1e-06,
"loss": 0.3088,
"step": 2335
},
{
"epoch": 0.8933078393881453,
"grad_norm": 3.054779597696523,
"learning_rate": 1e-06,
"loss": 0.0827,
"step": 2336
},
{
"epoch": 0.8936902485659656,
"grad_norm": 5.234545895647738,
"learning_rate": 1e-06,
"loss": 0.2402,
"step": 2337
},
{
"epoch": 0.8940726577437859,
"grad_norm": 5.033355651368099,
"learning_rate": 1e-06,
"loss": 0.1738,
"step": 2338
},
{
"epoch": 0.8944550669216061,
"grad_norm": 5.520875846651122,
"learning_rate": 1e-06,
"loss": 0.5797,
"step": 2339
},
{
"epoch": 0.8948374760994264,
"grad_norm": 3.5991780296358877,
"learning_rate": 1e-06,
"loss": 0.4866,
"step": 2340
},
{
"epoch": 0.8952198852772466,
"grad_norm": 2.7136232994721263,
"learning_rate": 1e-06,
"loss": 0.5725,
"step": 2341
},
{
"epoch": 0.8956022944550669,
"grad_norm": 2.5631846001892287,
"learning_rate": 1e-06,
"loss": 0.313,
"step": 2342
},
{
"epoch": 0.8959847036328872,
"grad_norm": 3.3386419621407724,
"learning_rate": 1e-06,
"loss": 0.3095,
"step": 2343
},
{
"epoch": 0.8963671128107075,
"grad_norm": 3.245795244247259,
"learning_rate": 1e-06,
"loss": 0.4238,
"step": 2344
},
{
"epoch": 0.8967495219885278,
"grad_norm": 2.2980130029879846,
"learning_rate": 1e-06,
"loss": 0.2659,
"step": 2345
},
{
"epoch": 0.897131931166348,
"grad_norm": 2.595626703634238,
"learning_rate": 1e-06,
"loss": 0.2533,
"step": 2346
},
{
"epoch": 0.8975143403441682,
"grad_norm": 2.7846215204267755,
"learning_rate": 1e-06,
"loss": 0.2174,
"step": 2347
},
{
"epoch": 0.8978967495219885,
"grad_norm": 3.3081022615874676,
"learning_rate": 1e-06,
"loss": 0.0999,
"step": 2348
},
{
"epoch": 0.8982791586998088,
"grad_norm": 3.0943818651222195,
"learning_rate": 1e-06,
"loss": 0.1473,
"step": 2349
},
{
"epoch": 0.8986615678776291,
"grad_norm": 4.190605947520891,
"learning_rate": 1e-06,
"loss": 0.1731,
"step": 2350
},
{
"epoch": 0.8990439770554494,
"grad_norm": 4.72621948006337,
"learning_rate": 1e-06,
"loss": 0.4197,
"step": 2351
},
{
"epoch": 0.8994263862332696,
"grad_norm": 3.5235686351792337,
"learning_rate": 1e-06,
"loss": 0.5667,
"step": 2352
},
{
"epoch": 0.8998087954110898,
"grad_norm": 2.469793075202874,
"learning_rate": 1e-06,
"loss": 0.2133,
"step": 2353
},
{
"epoch": 0.9001912045889101,
"grad_norm": 2.807314351874271,
"learning_rate": 1e-06,
"loss": 0.4659,
"step": 2354
},
{
"epoch": 0.9005736137667304,
"grad_norm": 3.795765153347245,
"learning_rate": 1e-06,
"loss": 0.4197,
"step": 2355
},
{
"epoch": 0.9009560229445507,
"grad_norm": 4.375152585229796,
"learning_rate": 1e-06,
"loss": 0.4471,
"step": 2356
},
{
"epoch": 0.901338432122371,
"grad_norm": 3.7510101229460644,
"learning_rate": 1e-06,
"loss": 0.2843,
"step": 2357
},
{
"epoch": 0.9017208413001913,
"grad_norm": 3.349252326998123,
"learning_rate": 1e-06,
"loss": 0.334,
"step": 2358
},
{
"epoch": 0.9021032504780114,
"grad_norm": 2.9929499439115674,
"learning_rate": 1e-06,
"loss": 0.2409,
"step": 2359
},
{
"epoch": 0.9024856596558317,
"grad_norm": 4.8274300578190426,
"learning_rate": 1e-06,
"loss": 0.1843,
"step": 2360
},
{
"epoch": 0.902868068833652,
"grad_norm": 3.04919407937432,
"learning_rate": 1e-06,
"loss": 0.12,
"step": 2361
},
{
"epoch": 0.9032504780114723,
"grad_norm": 3.614721920350499,
"learning_rate": 1e-06,
"loss": 0.1164,
"step": 2362
},
{
"epoch": 0.9036328871892926,
"grad_norm": 5.657648577523865,
"learning_rate": 1e-06,
"loss": 0.151,
"step": 2363
},
{
"epoch": 0.9040152963671129,
"grad_norm": 3.7782924214287825,
"learning_rate": 1e-06,
"loss": 0.5837,
"step": 2364
},
{
"epoch": 0.904397705544933,
"grad_norm": 3.29169842346582,
"learning_rate": 1e-06,
"loss": 0.4038,
"step": 2365
},
{
"epoch": 0.9047801147227533,
"grad_norm": 2.804095957650072,
"learning_rate": 1e-06,
"loss": 0.4941,
"step": 2366
},
{
"epoch": 0.9051625239005736,
"grad_norm": 2.6948585017325772,
"learning_rate": 1e-06,
"loss": 0.399,
"step": 2367
},
{
"epoch": 0.9055449330783939,
"grad_norm": 2.8048690694622755,
"learning_rate": 1e-06,
"loss": 0.3361,
"step": 2368
},
{
"epoch": 0.9059273422562142,
"grad_norm": 3.4714349201689294,
"learning_rate": 1e-06,
"loss": 0.4146,
"step": 2369
},
{
"epoch": 0.9063097514340345,
"grad_norm": 3.056515666125883,
"learning_rate": 1e-06,
"loss": 0.2881,
"step": 2370
},
{
"epoch": 0.9066921606118546,
"grad_norm": 3.8923882468271436,
"learning_rate": 1e-06,
"loss": 0.3238,
"step": 2371
},
{
"epoch": 0.9070745697896749,
"grad_norm": 3.9766025503723563,
"learning_rate": 1e-06,
"loss": 0.3097,
"step": 2372
},
{
"epoch": 0.9074569789674952,
"grad_norm": 3.991801204020184,
"learning_rate": 1e-06,
"loss": 0.2089,
"step": 2373
},
{
"epoch": 0.9078393881453155,
"grad_norm": 3.8195215822387363,
"learning_rate": 1e-06,
"loss": 0.1582,
"step": 2374
},
{
"epoch": 0.9082217973231358,
"grad_norm": 3.1414569113516273,
"learning_rate": 1e-06,
"loss": 0.0923,
"step": 2375
},
{
"epoch": 0.908604206500956,
"grad_norm": 4.024326974268517,
"learning_rate": 1e-06,
"loss": 0.3416,
"step": 2376
},
{
"epoch": 0.9089866156787763,
"grad_norm": 3.5697654121323223,
"learning_rate": 1e-06,
"loss": 0.6562,
"step": 2377
},
{
"epoch": 0.9093690248565965,
"grad_norm": 3.3032838261728106,
"learning_rate": 1e-06,
"loss": 0.3549,
"step": 2378
},
{
"epoch": 0.9097514340344168,
"grad_norm": 3.729689975157589,
"learning_rate": 1e-06,
"loss": 0.5737,
"step": 2379
},
{
"epoch": 0.9101338432122371,
"grad_norm": 3.084084874569733,
"learning_rate": 1e-06,
"loss": 0.4262,
"step": 2380
},
{
"epoch": 0.9105162523900574,
"grad_norm": 2.7431866599727837,
"learning_rate": 1e-06,
"loss": 0.2823,
"step": 2381
},
{
"epoch": 0.9108986615678776,
"grad_norm": 3.3325277944609093,
"learning_rate": 1e-06,
"loss": 0.1517,
"step": 2382
},
{
"epoch": 0.9112810707456979,
"grad_norm": 3.2485280738522864,
"learning_rate": 1e-06,
"loss": 0.3246,
"step": 2383
},
{
"epoch": 0.9116634799235181,
"grad_norm": 2.91043118195861,
"learning_rate": 1e-06,
"loss": 0.1999,
"step": 2384
},
{
"epoch": 0.9120458891013384,
"grad_norm": 2.892826705853231,
"learning_rate": 1e-06,
"loss": 0.1659,
"step": 2385
},
{
"epoch": 0.9124282982791587,
"grad_norm": 2.7669380648525954,
"learning_rate": 1e-06,
"loss": 0.1941,
"step": 2386
},
{
"epoch": 0.912810707456979,
"grad_norm": 3.397510761752007,
"learning_rate": 1e-06,
"loss": 0.0964,
"step": 2387
},
{
"epoch": 0.9131931166347992,
"grad_norm": 6.884696918983418,
"learning_rate": 1e-06,
"loss": 0.1733,
"step": 2388
},
{
"epoch": 0.9135755258126195,
"grad_norm": 4.374020276044012,
"learning_rate": 1e-06,
"loss": 0.4105,
"step": 2389
},
{
"epoch": 0.9139579349904398,
"grad_norm": 2.4249034036490626,
"learning_rate": 1e-06,
"loss": 0.3799,
"step": 2390
},
{
"epoch": 0.91434034416826,
"grad_norm": 1.7467237867717293,
"learning_rate": 1e-06,
"loss": 0.5461,
"step": 2391
},
{
"epoch": 0.9147227533460803,
"grad_norm": 3.769540990312869,
"learning_rate": 1e-06,
"loss": 0.3656,
"step": 2392
},
{
"epoch": 0.9151051625239006,
"grad_norm": 3.051950618909361,
"learning_rate": 1e-06,
"loss": 0.2499,
"step": 2393
},
{
"epoch": 0.9154875717017208,
"grad_norm": 2.3185247347830433,
"learning_rate": 1e-06,
"loss": 0.2253,
"step": 2394
},
{
"epoch": 0.9158699808795411,
"grad_norm": 3.6773879201524555,
"learning_rate": 1e-06,
"loss": 0.2363,
"step": 2395
},
{
"epoch": 0.9162523900573614,
"grad_norm": 2.7726956946995234,
"learning_rate": 1e-06,
"loss": 0.2691,
"step": 2396
},
{
"epoch": 0.9166347992351817,
"grad_norm": 4.293276636053358,
"learning_rate": 1e-06,
"loss": 0.1415,
"step": 2397
},
{
"epoch": 0.9170172084130019,
"grad_norm": 3.2937914272070086,
"learning_rate": 1e-06,
"loss": 0.1545,
"step": 2398
},
{
"epoch": 0.9173996175908222,
"grad_norm": 5.068736061989545,
"learning_rate": 1e-06,
"loss": 0.1117,
"step": 2399
},
{
"epoch": 0.9177820267686424,
"grad_norm": 4.144171811551991,
"learning_rate": 1e-06,
"loss": 0.211,
"step": 2400
},
{
"epoch": 0.9181644359464627,
"grad_norm": 4.271453564275025,
"learning_rate": 1e-06,
"loss": 0.4232,
"step": 2401
},
{
"epoch": 0.918546845124283,
"grad_norm": 4.142120897238968,
"learning_rate": 1e-06,
"loss": 0.3927,
"step": 2402
},
{
"epoch": 0.9189292543021033,
"grad_norm": 2.4200139924898028,
"learning_rate": 1e-06,
"loss": 0.3811,
"step": 2403
},
{
"epoch": 0.9193116634799235,
"grad_norm": 4.018243907902251,
"learning_rate": 1e-06,
"loss": 0.4066,
"step": 2404
},
{
"epoch": 0.9196940726577438,
"grad_norm": 2.7654656833084443,
"learning_rate": 1e-06,
"loss": 0.3045,
"step": 2405
},
{
"epoch": 0.920076481835564,
"grad_norm": 3.8460711162179333,
"learning_rate": 1e-06,
"loss": 0.276,
"step": 2406
},
{
"epoch": 0.9204588910133843,
"grad_norm": 3.2811800994693674,
"learning_rate": 1e-06,
"loss": 0.3471,
"step": 2407
},
{
"epoch": 0.9208413001912046,
"grad_norm": 2.816157188460834,
"learning_rate": 1e-06,
"loss": 0.271,
"step": 2408
},
{
"epoch": 0.9212237093690249,
"grad_norm": 4.139433145830336,
"learning_rate": 1e-06,
"loss": 0.321,
"step": 2409
},
{
"epoch": 0.9216061185468452,
"grad_norm": 3.250587116908922,
"learning_rate": 1e-06,
"loss": 0.1285,
"step": 2410
},
{
"epoch": 0.9219885277246654,
"grad_norm": 2.511674990354371,
"learning_rate": 1e-06,
"loss": 0.1159,
"step": 2411
},
{
"epoch": 0.9223709369024856,
"grad_norm": 2.8328083524342222,
"learning_rate": 1e-06,
"loss": 0.0856,
"step": 2412
},
{
"epoch": 0.9227533460803059,
"grad_norm": 4.623206950914736,
"learning_rate": 1e-06,
"loss": 0.2897,
"step": 2413
},
{
"epoch": 0.9231357552581262,
"grad_norm": 5.678655769427854,
"learning_rate": 1e-06,
"loss": 0.5478,
"step": 2414
},
{
"epoch": 0.9235181644359465,
"grad_norm": 3.1158680047222096,
"learning_rate": 1e-06,
"loss": 0.6466,
"step": 2415
},
{
"epoch": 0.9239005736137668,
"grad_norm": 2.4494728066125724,
"learning_rate": 1e-06,
"loss": 0.2526,
"step": 2416
},
{
"epoch": 0.924282982791587,
"grad_norm": 2.9669818632152016,
"learning_rate": 1e-06,
"loss": 0.3885,
"step": 2417
},
{
"epoch": 0.9246653919694072,
"grad_norm": 3.402103525083524,
"learning_rate": 1e-06,
"loss": 0.2849,
"step": 2418
},
{
"epoch": 0.9250478011472275,
"grad_norm": 3.453406274999731,
"learning_rate": 1e-06,
"loss": 0.3591,
"step": 2419
},
{
"epoch": 0.9254302103250478,
"grad_norm": 2.0307213902395898,
"learning_rate": 1e-06,
"loss": 0.1947,
"step": 2420
},
{
"epoch": 0.9258126195028681,
"grad_norm": 2.5965596802903117,
"learning_rate": 1e-06,
"loss": 0.1767,
"step": 2421
},
{
"epoch": 0.9261950286806884,
"grad_norm": 2.2293907005413836,
"learning_rate": 1e-06,
"loss": 0.0941,
"step": 2422
},
{
"epoch": 0.9265774378585087,
"grad_norm": 3.9204266674070825,
"learning_rate": 1e-06,
"loss": 0.2149,
"step": 2423
},
{
"epoch": 0.9269598470363288,
"grad_norm": 4.919483197059165,
"learning_rate": 1e-06,
"loss": 0.1295,
"step": 2424
},
{
"epoch": 0.9273422562141491,
"grad_norm": 7.121928573614454,
"learning_rate": 1e-06,
"loss": 0.306,
"step": 2425
},
{
"epoch": 0.9277246653919694,
"grad_norm": 6.867140560402005,
"learning_rate": 1e-06,
"loss": 0.3018,
"step": 2426
},
{
"epoch": 0.9281070745697897,
"grad_norm": 3.502978419998348,
"learning_rate": 1e-06,
"loss": 0.4303,
"step": 2427
},
{
"epoch": 0.92848948374761,
"grad_norm": 2.5172368921493087,
"learning_rate": 1e-06,
"loss": 0.3832,
"step": 2428
},
{
"epoch": 0.9288718929254303,
"grad_norm": 2.401462399463372,
"learning_rate": 1e-06,
"loss": 0.2555,
"step": 2429
},
{
"epoch": 0.9292543021032504,
"grad_norm": 3.824808357773478,
"learning_rate": 1e-06,
"loss": 0.1962,
"step": 2430
},
{
"epoch": 0.9296367112810707,
"grad_norm": 2.7748829688060845,
"learning_rate": 1e-06,
"loss": 0.3494,
"step": 2431
},
{
"epoch": 0.930019120458891,
"grad_norm": 2.8105509468351326,
"learning_rate": 1e-06,
"loss": 0.2075,
"step": 2432
},
{
"epoch": 0.9304015296367113,
"grad_norm": 2.543391648153547,
"learning_rate": 1e-06,
"loss": 0.1866,
"step": 2433
},
{
"epoch": 0.9307839388145316,
"grad_norm": 3.660156771111222,
"learning_rate": 1e-06,
"loss": 0.3947,
"step": 2434
},
{
"epoch": 0.9311663479923518,
"grad_norm": 2.6722867771087198,
"learning_rate": 1e-06,
"loss": 0.1001,
"step": 2435
},
{
"epoch": 0.931548757170172,
"grad_norm": 4.882081585919296,
"learning_rate": 1e-06,
"loss": 0.1693,
"step": 2436
},
{
"epoch": 0.9319311663479923,
"grad_norm": 3.8891033582736587,
"learning_rate": 1e-06,
"loss": 0.0689,
"step": 2437
},
{
"epoch": 0.9323135755258126,
"grad_norm": 5.030727762164894,
"learning_rate": 1e-06,
"loss": 0.1524,
"step": 2438
},
{
"epoch": 0.9326959847036329,
"grad_norm": 2.806366824158863,
"learning_rate": 1e-06,
"loss": 0.3798,
"step": 2439
},
{
"epoch": 0.9330783938814532,
"grad_norm": 3.993293385080311,
"learning_rate": 1e-06,
"loss": 0.4549,
"step": 2440
},
{
"epoch": 0.9334608030592734,
"grad_norm": 2.6540438297737485,
"learning_rate": 1e-06,
"loss": 0.4144,
"step": 2441
},
{
"epoch": 0.9338432122370937,
"grad_norm": 3.9939570079279467,
"learning_rate": 1e-06,
"loss": 0.2668,
"step": 2442
},
{
"epoch": 0.9342256214149139,
"grad_norm": 3.8146261321492796,
"learning_rate": 1e-06,
"loss": 0.4552,
"step": 2443
},
{
"epoch": 0.9346080305927342,
"grad_norm": 3.880128973164252,
"learning_rate": 1e-06,
"loss": 0.1813,
"step": 2444
},
{
"epoch": 0.9349904397705545,
"grad_norm": 4.808416243994762,
"learning_rate": 1e-06,
"loss": 0.2827,
"step": 2445
},
{
"epoch": 0.9353728489483748,
"grad_norm": 3.4106847609138664,
"learning_rate": 1e-06,
"loss": 0.1591,
"step": 2446
},
{
"epoch": 0.935755258126195,
"grad_norm": 2.4377755229460027,
"learning_rate": 1e-06,
"loss": 0.1474,
"step": 2447
},
{
"epoch": 0.9361376673040153,
"grad_norm": 3.229732406444275,
"learning_rate": 1e-06,
"loss": 0.1188,
"step": 2448
},
{
"epoch": 0.9365200764818356,
"grad_norm": 3.598057296881959,
"learning_rate": 1e-06,
"loss": 0.1297,
"step": 2449
},
{
"epoch": 0.9369024856596558,
"grad_norm": 4.248837031398682,
"learning_rate": 1e-06,
"loss": 0.1598,
"step": 2450
},
{
"epoch": 0.9372848948374761,
"grad_norm": 5.293401338699338,
"learning_rate": 1e-06,
"loss": 0.5099,
"step": 2451
},
{
"epoch": 0.9376673040152964,
"grad_norm": 4.239485131078506,
"learning_rate": 1e-06,
"loss": 0.4603,
"step": 2452
},
{
"epoch": 0.9380497131931166,
"grad_norm": 3.178237621971313,
"learning_rate": 1e-06,
"loss": 0.3857,
"step": 2453
},
{
"epoch": 0.9384321223709369,
"grad_norm": 2.820377761212562,
"learning_rate": 1e-06,
"loss": 0.4409,
"step": 2454
},
{
"epoch": 0.9388145315487572,
"grad_norm": 4.224050173562621,
"learning_rate": 1e-06,
"loss": 0.4657,
"step": 2455
},
{
"epoch": 0.9391969407265774,
"grad_norm": 3.480768501226541,
"learning_rate": 1e-06,
"loss": 0.2117,
"step": 2456
},
{
"epoch": 0.9395793499043977,
"grad_norm": 3.3157551444179916,
"learning_rate": 1e-06,
"loss": 0.2881,
"step": 2457
},
{
"epoch": 0.939961759082218,
"grad_norm": 2.677380605933176,
"learning_rate": 1e-06,
"loss": 0.2783,
"step": 2458
},
{
"epoch": 0.9403441682600382,
"grad_norm": 3.4601604889710007,
"learning_rate": 1e-06,
"loss": 0.1185,
"step": 2459
},
{
"epoch": 0.9407265774378585,
"grad_norm": 2.7606470167868156,
"learning_rate": 1e-06,
"loss": 0.1084,
"step": 2460
},
{
"epoch": 0.9411089866156788,
"grad_norm": 4.2092945524484815,
"learning_rate": 1e-06,
"loss": 0.1193,
"step": 2461
},
{
"epoch": 0.941491395793499,
"grad_norm": 4.857380252133691,
"learning_rate": 1e-06,
"loss": 0.2438,
"step": 2462
},
{
"epoch": 0.9418738049713193,
"grad_norm": 4.812435991926709,
"learning_rate": 1e-06,
"loss": 0.1943,
"step": 2463
},
{
"epoch": 0.9422562141491396,
"grad_norm": 3.0303294563586163,
"learning_rate": 1e-06,
"loss": 0.3405,
"step": 2464
},
{
"epoch": 0.9426386233269598,
"grad_norm": 3.4931876779237387,
"learning_rate": 1e-06,
"loss": 0.3735,
"step": 2465
},
{
"epoch": 0.9430210325047801,
"grad_norm": 3.105325947483179,
"learning_rate": 1e-06,
"loss": 0.193,
"step": 2466
},
{
"epoch": 0.9434034416826004,
"grad_norm": 3.3599543581996127,
"learning_rate": 1e-06,
"loss": 0.4249,
"step": 2467
},
{
"epoch": 0.9437858508604207,
"grad_norm": 3.9436596117981617,
"learning_rate": 1e-06,
"loss": 0.4425,
"step": 2468
},
{
"epoch": 0.944168260038241,
"grad_norm": 2.914177531817628,
"learning_rate": 1e-06,
"loss": 0.4259,
"step": 2469
},
{
"epoch": 0.9445506692160612,
"grad_norm": 3.3477942706469674,
"learning_rate": 1e-06,
"loss": 0.1776,
"step": 2470
},
{
"epoch": 0.9449330783938814,
"grad_norm": 3.404480500775937,
"learning_rate": 1e-06,
"loss": 0.3247,
"step": 2471
},
{
"epoch": 0.9453154875717017,
"grad_norm": 4.0445641949383795,
"learning_rate": 1e-06,
"loss": 0.1704,
"step": 2472
},
{
"epoch": 0.945697896749522,
"grad_norm": 4.140376706146293,
"learning_rate": 1e-06,
"loss": 0.1252,
"step": 2473
},
{
"epoch": 0.9460803059273423,
"grad_norm": 4.425210618001349,
"learning_rate": 1e-06,
"loss": 0.166,
"step": 2474
},
{
"epoch": 0.9464627151051626,
"grad_norm": 3.84246047804393,
"learning_rate": 1e-06,
"loss": 0.1434,
"step": 2475
},
{
"epoch": 0.9468451242829828,
"grad_norm": 4.633907503919647,
"learning_rate": 1e-06,
"loss": 0.5574,
"step": 2476
},
{
"epoch": 0.947227533460803,
"grad_norm": 3.4089993921191524,
"learning_rate": 1e-06,
"loss": 0.3722,
"step": 2477
},
{
"epoch": 0.9476099426386233,
"grad_norm": 2.981419562869889,
"learning_rate": 1e-06,
"loss": 0.3436,
"step": 2478
},
{
"epoch": 0.9479923518164436,
"grad_norm": 3.10857530594632,
"learning_rate": 1e-06,
"loss": 0.4303,
"step": 2479
},
{
"epoch": 0.9483747609942639,
"grad_norm": 2.475556467095296,
"learning_rate": 1e-06,
"loss": 0.4754,
"step": 2480
},
{
"epoch": 0.9487571701720842,
"grad_norm": 3.4656399367950486,
"learning_rate": 1e-06,
"loss": 0.3549,
"step": 2481
},
{
"epoch": 0.9491395793499044,
"grad_norm": 2.481484416967314,
"learning_rate": 1e-06,
"loss": 0.2656,
"step": 2482
},
{
"epoch": 0.9495219885277246,
"grad_norm": 3.2459105659146714,
"learning_rate": 1e-06,
"loss": 0.1877,
"step": 2483
},
{
"epoch": 0.9499043977055449,
"grad_norm": 2.9867060480189456,
"learning_rate": 1e-06,
"loss": 0.2577,
"step": 2484
},
{
"epoch": 0.9502868068833652,
"grad_norm": 2.9665059492603536,
"learning_rate": 1e-06,
"loss": 0.1622,
"step": 2485
},
{
"epoch": 0.9506692160611855,
"grad_norm": 3.2158824406103794,
"learning_rate": 1e-06,
"loss": 0.1046,
"step": 2486
},
{
"epoch": 0.9510516252390058,
"grad_norm": 7.309422032668531,
"learning_rate": 1e-06,
"loss": 0.2142,
"step": 2487
},
{
"epoch": 0.9514340344168261,
"grad_norm": 6.018173985632012,
"learning_rate": 1e-06,
"loss": 0.129,
"step": 2488
},
{
"epoch": 0.9518164435946462,
"grad_norm": 3.9050742859557754,
"learning_rate": 1e-06,
"loss": 0.2745,
"step": 2489
},
{
"epoch": 0.9521988527724665,
"grad_norm": 3.0204257207766267,
"learning_rate": 1e-06,
"loss": 0.5162,
"step": 2490
},
{
"epoch": 0.9525812619502868,
"grad_norm": 2.9900339210142612,
"learning_rate": 1e-06,
"loss": 0.2981,
"step": 2491
},
{
"epoch": 0.9529636711281071,
"grad_norm": 2.2315165769186533,
"learning_rate": 1e-06,
"loss": 0.2601,
"step": 2492
},
{
"epoch": 0.9533460803059274,
"grad_norm": 4.215404625202046,
"learning_rate": 1e-06,
"loss": 0.5005,
"step": 2493
},
{
"epoch": 0.9537284894837476,
"grad_norm": 4.286058607347795,
"learning_rate": 1e-06,
"loss": 0.3125,
"step": 2494
},
{
"epoch": 0.9541108986615678,
"grad_norm": 2.9477273365773953,
"learning_rate": 1e-06,
"loss": 0.1585,
"step": 2495
},
{
"epoch": 0.9544933078393881,
"grad_norm": 2.871610052694963,
"learning_rate": 1e-06,
"loss": 0.3119,
"step": 2496
},
{
"epoch": 0.9548757170172084,
"grad_norm": 4.013999045743957,
"learning_rate": 1e-06,
"loss": 0.1336,
"step": 2497
},
{
"epoch": 0.9552581261950287,
"grad_norm": 5.1393778021386245,
"learning_rate": 1e-06,
"loss": 0.1823,
"step": 2498
},
{
"epoch": 0.955640535372849,
"grad_norm": 4.134935378246604,
"learning_rate": 1e-06,
"loss": 0.1288,
"step": 2499
},
{
"epoch": 0.9560229445506692,
"grad_norm": 3.8629845176424045,
"learning_rate": 1e-06,
"loss": 0.2066,
"step": 2500
},
{
"epoch": 0.9560229445506692,
"eval_runtime": 799.3556,
"eval_samples_per_second": 1.919,
"eval_steps_per_second": 0.48,
"step": 2500
},
{
"epoch": 0.9564053537284894,
"grad_norm": 5.8565165553706136,
"learning_rate": 1e-06,
"loss": 0.2651,
"step": 2501
},
{
"epoch": 0.9567877629063097,
"grad_norm": 4.526794347596287,
"learning_rate": 1e-06,
"loss": 0.5482,
"step": 2502
},
{
"epoch": 0.95717017208413,
"grad_norm": 3.848471618079834,
"learning_rate": 1e-06,
"loss": 0.3268,
"step": 2503
},
{
"epoch": 0.9575525812619503,
"grad_norm": 2.5777081701684113,
"learning_rate": 1e-06,
"loss": 0.375,
"step": 2504
},
{
"epoch": 0.9579349904397706,
"grad_norm": 3.2761608075834734,
"learning_rate": 1e-06,
"loss": 0.4698,
"step": 2505
},
{
"epoch": 0.9583173996175908,
"grad_norm": 3.1059146201145658,
"learning_rate": 1e-06,
"loss": 0.3876,
"step": 2506
},
{
"epoch": 0.9586998087954111,
"grad_norm": 4.5630668392531994,
"learning_rate": 1e-06,
"loss": 0.293,
"step": 2507
},
{
"epoch": 0.9590822179732313,
"grad_norm": 2.5173547141041355,
"learning_rate": 1e-06,
"loss": 0.4098,
"step": 2508
},
{
"epoch": 0.9594646271510516,
"grad_norm": 4.096442116360893,
"learning_rate": 1e-06,
"loss": 0.3183,
"step": 2509
},
{
"epoch": 0.9598470363288719,
"grad_norm": 3.040329539105895,
"learning_rate": 1e-06,
"loss": 0.1837,
"step": 2510
},
{
"epoch": 0.9602294455066922,
"grad_norm": 4.060086046631855,
"learning_rate": 1e-06,
"loss": 0.1759,
"step": 2511
},
{
"epoch": 0.9606118546845124,
"grad_norm": 3.202650974833437,
"learning_rate": 1e-06,
"loss": 0.1191,
"step": 2512
},
{
"epoch": 0.9609942638623327,
"grad_norm": 5.244336297559902,
"learning_rate": 1e-06,
"loss": 0.221,
"step": 2513
},
{
"epoch": 0.961376673040153,
"grad_norm": 4.087860530153794,
"learning_rate": 1e-06,
"loss": 0.4887,
"step": 2514
},
{
"epoch": 0.9617590822179732,
"grad_norm": 4.008707106538918,
"learning_rate": 1e-06,
"loss": 0.2974,
"step": 2515
},
{
"epoch": 0.9621414913957935,
"grad_norm": 2.8413734985847023,
"learning_rate": 1e-06,
"loss": 0.2436,
"step": 2516
},
{
"epoch": 0.9625239005736138,
"grad_norm": 2.1263401069363788,
"learning_rate": 1e-06,
"loss": 0.4188,
"step": 2517
},
{
"epoch": 0.962906309751434,
"grad_norm": 2.3591884514025065,
"learning_rate": 1e-06,
"loss": 0.3366,
"step": 2518
},
{
"epoch": 0.9632887189292543,
"grad_norm": 3.4527239674509396,
"learning_rate": 1e-06,
"loss": 0.3246,
"step": 2519
},
{
"epoch": 0.9636711281070746,
"grad_norm": 3.1959436814522637,
"learning_rate": 1e-06,
"loss": 0.4524,
"step": 2520
},
{
"epoch": 0.9640535372848948,
"grad_norm": 2.6434737870598064,
"learning_rate": 1e-06,
"loss": 0.1332,
"step": 2521
},
{
"epoch": 0.9644359464627151,
"grad_norm": 2.9888744687942252,
"learning_rate": 1e-06,
"loss": 0.2597,
"step": 2522
},
{
"epoch": 0.9648183556405354,
"grad_norm": 2.8214012855189923,
"learning_rate": 1e-06,
"loss": 0.095,
"step": 2523
},
{
"epoch": 0.9652007648183556,
"grad_norm": 3.6102344100802566,
"learning_rate": 1e-06,
"loss": 0.2253,
"step": 2524
},
{
"epoch": 0.9655831739961759,
"grad_norm": 3.7707971532501547,
"learning_rate": 1e-06,
"loss": 0.1292,
"step": 2525
},
{
"epoch": 0.9659655831739962,
"grad_norm": 6.197414892919504,
"learning_rate": 1e-06,
"loss": 0.4118,
"step": 2526
},
{
"epoch": 0.9663479923518165,
"grad_norm": 4.232979970486748,
"learning_rate": 1e-06,
"loss": 0.525,
"step": 2527
},
{
"epoch": 0.9667304015296367,
"grad_norm": 2.516070213963481,
"learning_rate": 1e-06,
"loss": 0.2949,
"step": 2528
},
{
"epoch": 0.967112810707457,
"grad_norm": 2.5454068311021585,
"learning_rate": 1e-06,
"loss": 0.3649,
"step": 2529
},
{
"epoch": 0.9674952198852772,
"grad_norm": 3.0030260718779824,
"learning_rate": 1e-06,
"loss": 0.3655,
"step": 2530
},
{
"epoch": 0.9678776290630975,
"grad_norm": 2.587838203331088,
"learning_rate": 1e-06,
"loss": 0.1599,
"step": 2531
},
{
"epoch": 0.9682600382409178,
"grad_norm": 2.7870652342992006,
"learning_rate": 1e-06,
"loss": 0.2055,
"step": 2532
},
{
"epoch": 0.9686424474187381,
"grad_norm": 2.2474966958502223,
"learning_rate": 1e-06,
"loss": 0.2339,
"step": 2533
},
{
"epoch": 0.9690248565965583,
"grad_norm": 3.6074896838595163,
"learning_rate": 1e-06,
"loss": 0.2102,
"step": 2534
},
{
"epoch": 0.9694072657743786,
"grad_norm": 4.83622817242849,
"learning_rate": 1e-06,
"loss": 0.3518,
"step": 2535
},
{
"epoch": 0.9697896749521988,
"grad_norm": 4.1927200348811215,
"learning_rate": 1e-06,
"loss": 0.1836,
"step": 2536
},
{
"epoch": 0.9701720841300191,
"grad_norm": 2.8666457826016307,
"learning_rate": 1e-06,
"loss": 0.1056,
"step": 2537
},
{
"epoch": 0.9705544933078394,
"grad_norm": 4.817647979633222,
"learning_rate": 1e-06,
"loss": 0.16,
"step": 2538
},
{
"epoch": 0.9709369024856597,
"grad_norm": 4.854718396616762,
"learning_rate": 1e-06,
"loss": 0.3504,
"step": 2539
},
{
"epoch": 0.97131931166348,
"grad_norm": 4.5204929393590385,
"learning_rate": 1e-06,
"loss": 0.464,
"step": 2540
},
{
"epoch": 0.9717017208413002,
"grad_norm": 2.168201770803353,
"learning_rate": 1e-06,
"loss": 0.3012,
"step": 2541
},
{
"epoch": 0.9720841300191204,
"grad_norm": 2.494170162590611,
"learning_rate": 1e-06,
"loss": 0.3649,
"step": 2542
},
{
"epoch": 0.9724665391969407,
"grad_norm": 3.1576684464974107,
"learning_rate": 1e-06,
"loss": 0.3593,
"step": 2543
},
{
"epoch": 0.972848948374761,
"grad_norm": 2.856905678713925,
"learning_rate": 1e-06,
"loss": 0.4119,
"step": 2544
},
{
"epoch": 0.9732313575525813,
"grad_norm": 2.727520531175859,
"learning_rate": 1e-06,
"loss": 0.2192,
"step": 2545
},
{
"epoch": 0.9736137667304016,
"grad_norm": 4.18918313888206,
"learning_rate": 1e-06,
"loss": 0.2203,
"step": 2546
},
{
"epoch": 0.9739961759082219,
"grad_norm": 2.975542990234044,
"learning_rate": 1e-06,
"loss": 0.1945,
"step": 2547
},
{
"epoch": 0.974378585086042,
"grad_norm": 3.5107795384400453,
"learning_rate": 1e-06,
"loss": 0.2166,
"step": 2548
},
{
"epoch": 0.9747609942638623,
"grad_norm": 4.385758387404672,
"learning_rate": 1e-06,
"loss": 0.1169,
"step": 2549
},
{
"epoch": 0.9751434034416826,
"grad_norm": 4.209472401376355,
"learning_rate": 1e-06,
"loss": 0.0965,
"step": 2550
},
{
"epoch": 0.9755258126195029,
"grad_norm": 5.427274186872016,
"learning_rate": 1e-06,
"loss": 0.367,
"step": 2551
},
{
"epoch": 0.9759082217973232,
"grad_norm": 3.2644648339131286,
"learning_rate": 1e-06,
"loss": 0.2814,
"step": 2552
},
{
"epoch": 0.9762906309751435,
"grad_norm": 2.6137838065735473,
"learning_rate": 1e-06,
"loss": 0.4892,
"step": 2553
},
{
"epoch": 0.9766730401529636,
"grad_norm": 4.326270632337801,
"learning_rate": 1e-06,
"loss": 0.2991,
"step": 2554
},
{
"epoch": 0.9770554493307839,
"grad_norm": 3.050565548352456,
"learning_rate": 1e-06,
"loss": 0.1965,
"step": 2555
},
{
"epoch": 0.9774378585086042,
"grad_norm": 2.7712793684363803,
"learning_rate": 1e-06,
"loss": 0.4099,
"step": 2556
},
{
"epoch": 0.9778202676864245,
"grad_norm": 2.672851116499552,
"learning_rate": 1e-06,
"loss": 0.3241,
"step": 2557
},
{
"epoch": 0.9782026768642448,
"grad_norm": 3.88321062038747,
"learning_rate": 1e-06,
"loss": 0.2394,
"step": 2558
},
{
"epoch": 0.978585086042065,
"grad_norm": 2.5404347628870267,
"learning_rate": 1e-06,
"loss": 0.1967,
"step": 2559
},
{
"epoch": 0.9789674952198852,
"grad_norm": 4.196580748422142,
"learning_rate": 1e-06,
"loss": 0.1953,
"step": 2560
},
{
"epoch": 0.9793499043977055,
"grad_norm": 4.321526410325028,
"learning_rate": 1e-06,
"loss": 0.1945,
"step": 2561
},
{
"epoch": 0.9797323135755258,
"grad_norm": 3.620238498397353,
"learning_rate": 1e-06,
"loss": 0.128,
"step": 2562
},
{
"epoch": 0.9801147227533461,
"grad_norm": 5.664534261064663,
"learning_rate": 1e-06,
"loss": 0.3241,
"step": 2563
},
{
"epoch": 0.9804971319311664,
"grad_norm": 6.847996060806995,
"learning_rate": 1e-06,
"loss": 0.5073,
"step": 2564
},
{
"epoch": 0.9808795411089866,
"grad_norm": 3.571970882113405,
"learning_rate": 1e-06,
"loss": 0.513,
"step": 2565
},
{
"epoch": 0.9812619502868068,
"grad_norm": 2.4624811084622946,
"learning_rate": 1e-06,
"loss": 0.2175,
"step": 2566
},
{
"epoch": 0.9816443594646271,
"grad_norm": 3.7807054363969903,
"learning_rate": 1e-06,
"loss": 0.3091,
"step": 2567
},
{
"epoch": 0.9820267686424474,
"grad_norm": 2.8580470936274107,
"learning_rate": 1e-06,
"loss": 0.3945,
"step": 2568
},
{
"epoch": 0.9824091778202677,
"grad_norm": 2.351385673895787,
"learning_rate": 1e-06,
"loss": 0.1745,
"step": 2569
},
{
"epoch": 0.982791586998088,
"grad_norm": 2.1665177171811596,
"learning_rate": 1e-06,
"loss": 0.3091,
"step": 2570
},
{
"epoch": 0.9831739961759082,
"grad_norm": 2.3260906983631826,
"learning_rate": 1e-06,
"loss": 0.107,
"step": 2571
},
{
"epoch": 0.9835564053537285,
"grad_norm": 3.1163465087244333,
"learning_rate": 1e-06,
"loss": 0.2531,
"step": 2572
},
{
"epoch": 0.9839388145315487,
"grad_norm": 5.852612634199883,
"learning_rate": 1e-06,
"loss": 0.2657,
"step": 2573
},
{
"epoch": 0.984321223709369,
"grad_norm": 3.8884425648894876,
"learning_rate": 1e-06,
"loss": 0.2904,
"step": 2574
},
{
"epoch": 0.9847036328871893,
"grad_norm": 4.19222505362827,
"learning_rate": 1e-06,
"loss": 0.1286,
"step": 2575
},
{
"epoch": 0.9850860420650096,
"grad_norm": 5.189746600012265,
"learning_rate": 1e-06,
"loss": 0.4261,
"step": 2576
},
{
"epoch": 0.9854684512428298,
"grad_norm": 4.696277385973995,
"learning_rate": 1e-06,
"loss": 0.5642,
"step": 2577
},
{
"epoch": 0.9858508604206501,
"grad_norm": 3.4502829338930416,
"learning_rate": 1e-06,
"loss": 0.5265,
"step": 2578
},
{
"epoch": 0.9862332695984704,
"grad_norm": 3.274044007990694,
"learning_rate": 1e-06,
"loss": 0.318,
"step": 2579
},
{
"epoch": 0.9866156787762906,
"grad_norm": 2.9082146023740947,
"learning_rate": 1e-06,
"loss": 0.2627,
"step": 2580
},
{
"epoch": 0.9869980879541109,
"grad_norm": 3.030790156456422,
"learning_rate": 1e-06,
"loss": 0.3913,
"step": 2581
},
{
"epoch": 0.9873804971319312,
"grad_norm": 3.283557189694305,
"learning_rate": 1e-06,
"loss": 0.2163,
"step": 2582
},
{
"epoch": 0.9877629063097514,
"grad_norm": 3.117959445603992,
"learning_rate": 1e-06,
"loss": 0.2683,
"step": 2583
},
{
"epoch": 0.9881453154875717,
"grad_norm": 3.148377524436508,
"learning_rate": 1e-06,
"loss": 0.1586,
"step": 2584
},
{
"epoch": 0.988527724665392,
"grad_norm": 4.6813955612878,
"learning_rate": 1e-06,
"loss": 0.1509,
"step": 2585
},
{
"epoch": 0.9889101338432122,
"grad_norm": 3.899554809200781,
"learning_rate": 1e-06,
"loss": 0.1506,
"step": 2586
},
{
"epoch": 0.9892925430210325,
"grad_norm": 2.4800554066283054,
"learning_rate": 1e-06,
"loss": 0.0729,
"step": 2587
},
{
"epoch": 0.9896749521988528,
"grad_norm": 7.238309181847078,
"learning_rate": 1e-06,
"loss": 0.2698,
"step": 2588
},
{
"epoch": 0.990057361376673,
"grad_norm": 4.662581881297602,
"learning_rate": 1e-06,
"loss": 0.3281,
"step": 2589
},
{
"epoch": 0.9904397705544933,
"grad_norm": 2.739755187548956,
"learning_rate": 1e-06,
"loss": 0.5258,
"step": 2590
},
{
"epoch": 0.9908221797323136,
"grad_norm": 2.564721423929086,
"learning_rate": 1e-06,
"loss": 0.3786,
"step": 2591
},
{
"epoch": 0.9912045889101339,
"grad_norm": 2.5972957963319847,
"learning_rate": 1e-06,
"loss": 0.3551,
"step": 2592
},
{
"epoch": 0.9915869980879541,
"grad_norm": 3.8746655227526423,
"learning_rate": 1e-06,
"loss": 0.4603,
"step": 2593
},
{
"epoch": 0.9919694072657744,
"grad_norm": 2.229383749210781,
"learning_rate": 1e-06,
"loss": 0.2426,
"step": 2594
},
{
"epoch": 0.9923518164435946,
"grad_norm": 3.4840260147016933,
"learning_rate": 1e-06,
"loss": 0.3521,
"step": 2595
},
{
"epoch": 0.9927342256214149,
"grad_norm": 3.3125443365620715,
"learning_rate": 1e-06,
"loss": 0.3144,
"step": 2596
},
{
"epoch": 0.9931166347992352,
"grad_norm": 3.0452079711950035,
"learning_rate": 1e-06,
"loss": 0.1457,
"step": 2597
},
{
"epoch": 0.9934990439770555,
"grad_norm": 3.5410711049500967,
"learning_rate": 1e-06,
"loss": 0.2189,
"step": 2598
},
{
"epoch": 0.9938814531548757,
"grad_norm": 2.5759976944883465,
"learning_rate": 1e-06,
"loss": 0.063,
"step": 2599
},
{
"epoch": 0.994263862332696,
"grad_norm": 4.732681626241745,
"learning_rate": 1e-06,
"loss": 0.1459,
"step": 2600
},
{
"epoch": 0.9946462715105162,
"grad_norm": 5.7252164041855975,
"learning_rate": 1e-06,
"loss": 0.2892,
"step": 2601
},
{
"epoch": 0.9950286806883365,
"grad_norm": 3.9012106165781852,
"learning_rate": 1e-06,
"loss": 0.5871,
"step": 2602
},
{
"epoch": 0.9954110898661568,
"grad_norm": 4.077861675077326,
"learning_rate": 1e-06,
"loss": 0.3088,
"step": 2603
},
{
"epoch": 0.9957934990439771,
"grad_norm": 3.0754863447451473,
"learning_rate": 1e-06,
"loss": 0.3728,
"step": 2604
},
{
"epoch": 0.9961759082217974,
"grad_norm": 3.5602005347016674,
"learning_rate": 1e-06,
"loss": 0.3391,
"step": 2605
},
{
"epoch": 0.9965583173996176,
"grad_norm": 2.9985209633808587,
"learning_rate": 1e-06,
"loss": 0.3218,
"step": 2606
},
{
"epoch": 0.9969407265774378,
"grad_norm": 2.896954015417547,
"learning_rate": 1e-06,
"loss": 0.2512,
"step": 2607
},
{
"epoch": 0.9973231357552581,
"grad_norm": 2.8445699997682503,
"learning_rate": 1e-06,
"loss": 0.1779,
"step": 2608
},
{
"epoch": 0.9977055449330784,
"grad_norm": 3.448586359514333,
"learning_rate": 1e-06,
"loss": 0.2165,
"step": 2609
},
{
"epoch": 0.9980879541108987,
"grad_norm": 3.796851138444182,
"learning_rate": 1e-06,
"loss": 0.1629,
"step": 2610
},
{
"epoch": 0.998470363288719,
"grad_norm": 3.3308363781163437,
"learning_rate": 1e-06,
"loss": 0.1305,
"step": 2611
},
{
"epoch": 0.9988527724665393,
"grad_norm": 3.9866357709863354,
"learning_rate": 1e-06,
"loss": 0.1635,
"step": 2612
},
{
"epoch": 0.9992351816443594,
"grad_norm": 2.2663411028183136,
"learning_rate": 1e-06,
"loss": 0.2236,
"step": 2613
},
{
"epoch": 0.9996175908221797,
"grad_norm": 7.850867215017207,
"learning_rate": 1e-06,
"loss": 0.3879,
"step": 2614
},
{
"epoch": 1.0,
"grad_norm": 7.3600381928987115,
"learning_rate": 1e-06,
"loss": 0.2976,
"step": 2615
},
{
"epoch": 1.0003824091778202,
"grad_norm": 7.022242494277558,
"learning_rate": 1e-06,
"loss": 0.3882,
"step": 2616
},
{
"epoch": 1.0007648183556406,
"grad_norm": 3.2276398197207765,
"learning_rate": 1e-06,
"loss": 0.4019,
"step": 2617
},
{
"epoch": 1.0011472275334607,
"grad_norm": 4.711575339994919,
"learning_rate": 1e-06,
"loss": 0.384,
"step": 2618
},
{
"epoch": 1.0015296367112811,
"grad_norm": 3.964034996794371,
"learning_rate": 1e-06,
"loss": 0.3542,
"step": 2619
},
{
"epoch": 1.0019120458891013,
"grad_norm": 2.4735523305662164,
"learning_rate": 1e-06,
"loss": 0.367,
"step": 2620
},
{
"epoch": 1.0022944550669215,
"grad_norm": 2.984105921892883,
"learning_rate": 1e-06,
"loss": 0.2827,
"step": 2621
},
{
"epoch": 1.002676864244742,
"grad_norm": 2.690558711175746,
"learning_rate": 1e-06,
"loss": 0.2447,
"step": 2622
},
{
"epoch": 1.003059273422562,
"grad_norm": 2.195602777913192,
"learning_rate": 1e-06,
"loss": 0.2517,
"step": 2623
},
{
"epoch": 1.0034416826003825,
"grad_norm": 2.950277903769856,
"learning_rate": 1e-06,
"loss": 0.2829,
"step": 2624
},
{
"epoch": 1.0038240917782026,
"grad_norm": 2.2185522716692763,
"learning_rate": 1e-06,
"loss": 0.1924,
"step": 2625
},
{
"epoch": 1.004206500956023,
"grad_norm": 2.369061423893835,
"learning_rate": 1e-06,
"loss": 0.0956,
"step": 2626
},
{
"epoch": 1.0045889101338432,
"grad_norm": 3.1811822850303346,
"learning_rate": 1e-06,
"loss": 0.0704,
"step": 2627
},
{
"epoch": 1.0049713193116634,
"grad_norm": 4.695017368987337,
"learning_rate": 1e-06,
"loss": 0.2643,
"step": 2628
},
{
"epoch": 1.0053537284894838,
"grad_norm": 3.5123435749233964,
"learning_rate": 1e-06,
"loss": 0.3135,
"step": 2629
},
{
"epoch": 1.005736137667304,
"grad_norm": 4.023118682377531,
"learning_rate": 1e-06,
"loss": 0.4817,
"step": 2630
},
{
"epoch": 1.0061185468451244,
"grad_norm": 3.1225779493267387,
"learning_rate": 1e-06,
"loss": 0.5105,
"step": 2631
},
{
"epoch": 1.0065009560229445,
"grad_norm": 2.145657492040758,
"learning_rate": 1e-06,
"loss": 0.2007,
"step": 2632
},
{
"epoch": 1.0068833652007647,
"grad_norm": 3.395665794561428,
"learning_rate": 1e-06,
"loss": 0.2554,
"step": 2633
},
{
"epoch": 1.007265774378585,
"grad_norm": 2.738779237661642,
"learning_rate": 1e-06,
"loss": 0.2572,
"step": 2634
},
{
"epoch": 1.0076481835564053,
"grad_norm": 2.4865667883354314,
"learning_rate": 1e-06,
"loss": 0.3324,
"step": 2635
},
{
"epoch": 1.0080305927342257,
"grad_norm": 2.4270858328046465,
"learning_rate": 1e-06,
"loss": 0.199,
"step": 2636
},
{
"epoch": 1.0084130019120459,
"grad_norm": 2.83804453191572,
"learning_rate": 1e-06,
"loss": 0.2041,
"step": 2637
},
{
"epoch": 1.0087954110898663,
"grad_norm": 3.452598980500604,
"learning_rate": 1e-06,
"loss": 0.1231,
"step": 2638
},
{
"epoch": 1.0091778202676864,
"grad_norm": 3.1894734575285772,
"learning_rate": 1e-06,
"loss": 0.2265,
"step": 2639
},
{
"epoch": 1.0095602294455066,
"grad_norm": 4.684612147542248,
"learning_rate": 1e-06,
"loss": 0.1149,
"step": 2640
},
{
"epoch": 1.009942638623327,
"grad_norm": 6.372393505333487,
"learning_rate": 1e-06,
"loss": 0.3431,
"step": 2641
},
{
"epoch": 1.0103250478011472,
"grad_norm": 3.6672653807605475,
"learning_rate": 1e-06,
"loss": 0.4792,
"step": 2642
},
{
"epoch": 1.0107074569789676,
"grad_norm": 3.1671483192222225,
"learning_rate": 1e-06,
"loss": 0.1964,
"step": 2643
},
{
"epoch": 1.0110898661567878,
"grad_norm": 3.1009247169742444,
"learning_rate": 1e-06,
"loss": 0.3048,
"step": 2644
},
{
"epoch": 1.011472275334608,
"grad_norm": 2.561814565399431,
"learning_rate": 1e-06,
"loss": 0.4524,
"step": 2645
},
{
"epoch": 1.0118546845124283,
"grad_norm": 3.6103503738248652,
"learning_rate": 1e-06,
"loss": 0.4296,
"step": 2646
},
{
"epoch": 1.0122370936902485,
"grad_norm": 3.0457279491501494,
"learning_rate": 1e-06,
"loss": 0.2372,
"step": 2647
},
{
"epoch": 1.012619502868069,
"grad_norm": 2.6471071756222786,
"learning_rate": 1e-06,
"loss": 0.1504,
"step": 2648
},
{
"epoch": 1.013001912045889,
"grad_norm": 3.2796720070947787,
"learning_rate": 1e-06,
"loss": 0.3346,
"step": 2649
},
{
"epoch": 1.0133843212237095,
"grad_norm": 2.6879539993414476,
"learning_rate": 1e-06,
"loss": 0.1246,
"step": 2650
},
{
"epoch": 1.0137667304015296,
"grad_norm": 3.9806388541609574,
"learning_rate": 1e-06,
"loss": 0.2321,
"step": 2651
},
{
"epoch": 1.0141491395793498,
"grad_norm": 4.3342455612910635,
"learning_rate": 1e-06,
"loss": 0.1251,
"step": 2652
},
{
"epoch": 1.0145315487571702,
"grad_norm": 4.458886418246188,
"learning_rate": 1e-06,
"loss": 0.2768,
"step": 2653
},
{
"epoch": 1.0149139579349904,
"grad_norm": 3.829791768840152,
"learning_rate": 1e-06,
"loss": 0.7265,
"step": 2654
},
{
"epoch": 1.0152963671128108,
"grad_norm": 4.006958154705942,
"learning_rate": 1e-06,
"loss": 0.4493,
"step": 2655
},
{
"epoch": 1.015678776290631,
"grad_norm": 4.761598180256772,
"learning_rate": 1e-06,
"loss": 0.3316,
"step": 2656
},
{
"epoch": 1.0160611854684511,
"grad_norm": 2.4355635164452516,
"learning_rate": 1e-06,
"loss": 0.2961,
"step": 2657
},
{
"epoch": 1.0164435946462715,
"grad_norm": 3.4078103787408143,
"learning_rate": 1e-06,
"loss": 0.3313,
"step": 2658
},
{
"epoch": 1.0168260038240917,
"grad_norm": 2.336207345539462,
"learning_rate": 1e-06,
"loss": 0.266,
"step": 2659
},
{
"epoch": 1.0172084130019121,
"grad_norm": 2.483990525539346,
"learning_rate": 1e-06,
"loss": 0.2061,
"step": 2660
},
{
"epoch": 1.0175908221797323,
"grad_norm": 2.8088747183761535,
"learning_rate": 1e-06,
"loss": 0.2379,
"step": 2661
},
{
"epoch": 1.0179732313575527,
"grad_norm": 3.9472972749689803,
"learning_rate": 1e-06,
"loss": 0.1218,
"step": 2662
},
{
"epoch": 1.0183556405353729,
"grad_norm": 2.218365703129731,
"learning_rate": 1e-06,
"loss": 0.0802,
"step": 2663
},
{
"epoch": 1.018738049713193,
"grad_norm": 2.649953582645051,
"learning_rate": 1e-06,
"loss": 0.0778,
"step": 2664
},
{
"epoch": 1.0191204588910134,
"grad_norm": 4.448260732501134,
"learning_rate": 1e-06,
"loss": 0.1224,
"step": 2665
},
{
"epoch": 1.0195028680688336,
"grad_norm": 3.271824988098627,
"learning_rate": 1e-06,
"loss": 0.2896,
"step": 2666
},
{
"epoch": 1.019885277246654,
"grad_norm": 3.2242186671803053,
"learning_rate": 1e-06,
"loss": 0.3974,
"step": 2667
},
{
"epoch": 1.0202676864244742,
"grad_norm": 3.3364650478924704,
"learning_rate": 1e-06,
"loss": 0.3966,
"step": 2668
},
{
"epoch": 1.0206500956022944,
"grad_norm": 2.561851791645246,
"learning_rate": 1e-06,
"loss": 0.4702,
"step": 2669
},
{
"epoch": 1.0210325047801148,
"grad_norm": 2.296667180423467,
"learning_rate": 1e-06,
"loss": 0.1755,
"step": 2670
},
{
"epoch": 1.021414913957935,
"grad_norm": 2.5412507472527377,
"learning_rate": 1e-06,
"loss": 0.2024,
"step": 2671
},
{
"epoch": 1.0217973231357553,
"grad_norm": 3.1929982952396516,
"learning_rate": 1e-06,
"loss": 0.2564,
"step": 2672
},
{
"epoch": 1.0221797323135755,
"grad_norm": 3.12842570411532,
"learning_rate": 1e-06,
"loss": 0.1739,
"step": 2673
},
{
"epoch": 1.0225621414913957,
"grad_norm": 2.198754053160904,
"learning_rate": 1e-06,
"loss": 0.1205,
"step": 2674
},
{
"epoch": 1.022944550669216,
"grad_norm": 2.3468154188534847,
"learning_rate": 1e-06,
"loss": 0.099,
"step": 2675
},
{
"epoch": 1.0233269598470363,
"grad_norm": 2.9629685172276186,
"learning_rate": 1e-06,
"loss": 0.131,
"step": 2676
},
{
"epoch": 1.0237093690248567,
"grad_norm": 3.3928803005718886,
"learning_rate": 1e-06,
"loss": 0.0987,
"step": 2677
},
{
"epoch": 1.0240917782026768,
"grad_norm": 6.748226356603706,
"learning_rate": 1e-06,
"loss": 0.2138,
"step": 2678
},
{
"epoch": 1.0244741873804972,
"grad_norm": 4.177550609197598,
"learning_rate": 1e-06,
"loss": 0.4903,
"step": 2679
},
{
"epoch": 1.0248565965583174,
"grad_norm": 2.907307258180538,
"learning_rate": 1e-06,
"loss": 0.411,
"step": 2680
},
{
"epoch": 1.0252390057361376,
"grad_norm": 3.52547813581656,
"learning_rate": 1e-06,
"loss": 0.3489,
"step": 2681
},
{
"epoch": 1.025621414913958,
"grad_norm": 1.761864023922077,
"learning_rate": 1e-06,
"loss": 0.4432,
"step": 2682
},
{
"epoch": 1.0260038240917781,
"grad_norm": 4.062267120729046,
"learning_rate": 1e-06,
"loss": 0.5168,
"step": 2683
},
{
"epoch": 1.0263862332695985,
"grad_norm": 3.098152188290873,
"learning_rate": 1e-06,
"loss": 0.3486,
"step": 2684
},
{
"epoch": 1.0267686424474187,
"grad_norm": 2.888999358127268,
"learning_rate": 1e-06,
"loss": 0.3154,
"step": 2685
},
{
"epoch": 1.027151051625239,
"grad_norm": 2.1664383719064797,
"learning_rate": 1e-06,
"loss": 0.2393,
"step": 2686
},
{
"epoch": 1.0275334608030593,
"grad_norm": 2.937303577105565,
"learning_rate": 1e-06,
"loss": 0.1346,
"step": 2687
},
{
"epoch": 1.0279158699808795,
"grad_norm": 2.880227022759963,
"learning_rate": 1e-06,
"loss": 0.1251,
"step": 2688
},
{
"epoch": 1.0282982791586999,
"grad_norm": 5.532938898727842,
"learning_rate": 1e-06,
"loss": 0.1379,
"step": 2689
},
{
"epoch": 1.02868068833652,
"grad_norm": 4.385366963567743,
"learning_rate": 1e-06,
"loss": 0.1815,
"step": 2690
},
{
"epoch": 1.0290630975143404,
"grad_norm": 7.158234033869149,
"learning_rate": 1e-06,
"loss": 0.4647,
"step": 2691
},
{
"epoch": 1.0294455066921606,
"grad_norm": 3.0856505441641775,
"learning_rate": 1e-06,
"loss": 0.511,
"step": 2692
},
{
"epoch": 1.0298279158699808,
"grad_norm": 2.7459342552088613,
"learning_rate": 1e-06,
"loss": 0.396,
"step": 2693
},
{
"epoch": 1.0302103250478012,
"grad_norm": 2.322585408721161,
"learning_rate": 1e-06,
"loss": 0.2208,
"step": 2694
},
{
"epoch": 1.0305927342256214,
"grad_norm": 3.9542064770596155,
"learning_rate": 1e-06,
"loss": 0.2635,
"step": 2695
},
{
"epoch": 1.0309751434034418,
"grad_norm": 3.8928897501328565,
"learning_rate": 1e-06,
"loss": 0.3911,
"step": 2696
},
{
"epoch": 1.031357552581262,
"grad_norm": 2.718816822151605,
"learning_rate": 1e-06,
"loss": 0.2699,
"step": 2697
},
{
"epoch": 1.0317399617590821,
"grad_norm": 2.7116925070417888,
"learning_rate": 1e-06,
"loss": 0.1533,
"step": 2698
},
{
"epoch": 1.0321223709369025,
"grad_norm": 3.7386232098118484,
"learning_rate": 1e-06,
"loss": 0.2102,
"step": 2699
},
{
"epoch": 1.0325047801147227,
"grad_norm": 4.017373027397352,
"learning_rate": 1e-06,
"loss": 0.2086,
"step": 2700
},
{
"epoch": 1.032887189292543,
"grad_norm": 3.4537132072905083,
"learning_rate": 1e-06,
"loss": 0.0745,
"step": 2701
},
{
"epoch": 1.0332695984703633,
"grad_norm": 3.052307919169174,
"learning_rate": 1e-06,
"loss": 0.0833,
"step": 2702
},
{
"epoch": 1.0336520076481837,
"grad_norm": 5.773778758203473,
"learning_rate": 1e-06,
"loss": 0.2878,
"step": 2703
},
{
"epoch": 1.0340344168260038,
"grad_norm": 3.5658561558062556,
"learning_rate": 1e-06,
"loss": 0.3941,
"step": 2704
},
{
"epoch": 1.034416826003824,
"grad_norm": 3.8316777150803865,
"learning_rate": 1e-06,
"loss": 0.3779,
"step": 2705
},
{
"epoch": 1.0347992351816444,
"grad_norm": 3.0384828460897135,
"learning_rate": 1e-06,
"loss": 0.4879,
"step": 2706
},
{
"epoch": 1.0351816443594646,
"grad_norm": 2.343134786606256,
"learning_rate": 1e-06,
"loss": 0.176,
"step": 2707
},
{
"epoch": 1.035564053537285,
"grad_norm": 2.509989996354431,
"learning_rate": 1e-06,
"loss": 0.3084,
"step": 2708
},
{
"epoch": 1.0359464627151052,
"grad_norm": 4.5406688795706485,
"learning_rate": 1e-06,
"loss": 0.2898,
"step": 2709
},
{
"epoch": 1.0363288718929253,
"grad_norm": 2.519332617684988,
"learning_rate": 1e-06,
"loss": 0.3148,
"step": 2710
},
{
"epoch": 1.0367112810707457,
"grad_norm": 3.3118057963060408,
"learning_rate": 1e-06,
"loss": 0.1704,
"step": 2711
},
{
"epoch": 1.037093690248566,
"grad_norm": 3.1056092428420667,
"learning_rate": 1e-06,
"loss": 0.2319,
"step": 2712
},
{
"epoch": 1.0374760994263863,
"grad_norm": 2.8239605839709903,
"learning_rate": 1e-06,
"loss": 0.1413,
"step": 2713
},
{
"epoch": 1.0378585086042065,
"grad_norm": 3.6473681171255157,
"learning_rate": 1e-06,
"loss": 0.0879,
"step": 2714
},
{
"epoch": 1.0382409177820269,
"grad_norm": 4.901069773459302,
"learning_rate": 1e-06,
"loss": 0.1012,
"step": 2715
},
{
"epoch": 1.038623326959847,
"grad_norm": 5.90739502479764,
"learning_rate": 1e-06,
"loss": 0.3751,
"step": 2716
},
{
"epoch": 1.0390057361376672,
"grad_norm": 2.4297200780321946,
"learning_rate": 1e-06,
"loss": 0.3753,
"step": 2717
},
{
"epoch": 1.0393881453154876,
"grad_norm": 3.155282353796207,
"learning_rate": 1e-06,
"loss": 0.2652,
"step": 2718
},
{
"epoch": 1.0397705544933078,
"grad_norm": 1.8522253338930603,
"learning_rate": 1e-06,
"loss": 0.4514,
"step": 2719
},
{
"epoch": 1.0401529636711282,
"grad_norm": 3.8440830737692897,
"learning_rate": 1e-06,
"loss": 0.3758,
"step": 2720
},
{
"epoch": 1.0405353728489484,
"grad_norm": 3.6345675259188495,
"learning_rate": 1e-06,
"loss": 0.492,
"step": 2721
},
{
"epoch": 1.0409177820267685,
"grad_norm": 2.5366856174872265,
"learning_rate": 1e-06,
"loss": 0.3273,
"step": 2722
},
{
"epoch": 1.041300191204589,
"grad_norm": 2.9382334665978878,
"learning_rate": 1e-06,
"loss": 0.1969,
"step": 2723
},
{
"epoch": 1.0416826003824091,
"grad_norm": 2.463847447156355,
"learning_rate": 1e-06,
"loss": 0.1471,
"step": 2724
},
{
"epoch": 1.0420650095602295,
"grad_norm": 2.7102589926182095,
"learning_rate": 1e-06,
"loss": 0.0894,
"step": 2725
},
{
"epoch": 1.0424474187380497,
"grad_norm": 1.9479837117072258,
"learning_rate": 1e-06,
"loss": 0.0475,
"step": 2726
},
{
"epoch": 1.0428298279158699,
"grad_norm": 3.6407804660592453,
"learning_rate": 1e-06,
"loss": 0.0815,
"step": 2727
},
{
"epoch": 1.0432122370936903,
"grad_norm": 5.984271414634673,
"learning_rate": 1e-06,
"loss": 0.1927,
"step": 2728
},
{
"epoch": 1.0435946462715104,
"grad_norm": 3.93073151145546,
"learning_rate": 1e-06,
"loss": 0.5152,
"step": 2729
},
{
"epoch": 1.0439770554493308,
"grad_norm": 2.9375285695086757,
"learning_rate": 1e-06,
"loss": 0.434,
"step": 2730
},
{
"epoch": 1.044359464627151,
"grad_norm": 3.359699038468975,
"learning_rate": 1e-06,
"loss": 0.2864,
"step": 2731
},
{
"epoch": 1.0447418738049714,
"grad_norm": 3.754311181496516,
"learning_rate": 1e-06,
"loss": 0.2893,
"step": 2732
},
{
"epoch": 1.0451242829827916,
"grad_norm": 1.943514563723544,
"learning_rate": 1e-06,
"loss": 0.1743,
"step": 2733
},
{
"epoch": 1.0455066921606118,
"grad_norm": 3.242376188739394,
"learning_rate": 1e-06,
"loss": 0.2706,
"step": 2734
},
{
"epoch": 1.0458891013384322,
"grad_norm": 3.579687052426587,
"learning_rate": 1e-06,
"loss": 0.1136,
"step": 2735
},
{
"epoch": 1.0462715105162523,
"grad_norm": 2.2774925117913933,
"learning_rate": 1e-06,
"loss": 0.1565,
"step": 2736
},
{
"epoch": 1.0466539196940727,
"grad_norm": 2.907590495412597,
"learning_rate": 1e-06,
"loss": 0.1114,
"step": 2737
},
{
"epoch": 1.047036328871893,
"grad_norm": 3.855675039598945,
"learning_rate": 1e-06,
"loss": 0.1473,
"step": 2738
},
{
"epoch": 1.0474187380497133,
"grad_norm": 5.640318463309826,
"learning_rate": 1e-06,
"loss": 0.1222,
"step": 2739
},
{
"epoch": 1.0478011472275335,
"grad_norm": 3.7550896755030148,
"learning_rate": 1e-06,
"loss": 0.0826,
"step": 2740
},
{
"epoch": 1.0481835564053537,
"grad_norm": 8.14950996283073,
"learning_rate": 1e-06,
"loss": 0.4166,
"step": 2741
},
{
"epoch": 1.048565965583174,
"grad_norm": 3.7264255842412437,
"learning_rate": 1e-06,
"loss": 0.2951,
"step": 2742
},
{
"epoch": 1.0489483747609942,
"grad_norm": 3.1613406683671674,
"learning_rate": 1e-06,
"loss": 0.3623,
"step": 2743
},
{
"epoch": 1.0493307839388146,
"grad_norm": 2.42187224357202,
"learning_rate": 1e-06,
"loss": 0.5783,
"step": 2744
},
{
"epoch": 1.0497131931166348,
"grad_norm": 3.827291717144456,
"learning_rate": 1e-06,
"loss": 0.4027,
"step": 2745
},
{
"epoch": 1.050095602294455,
"grad_norm": 3.5942520454349443,
"learning_rate": 1e-06,
"loss": 0.2248,
"step": 2746
},
{
"epoch": 1.0504780114722754,
"grad_norm": 3.0847969361776753,
"learning_rate": 1e-06,
"loss": 0.2589,
"step": 2747
},
{
"epoch": 1.0508604206500956,
"grad_norm": 2.6449802755619127,
"learning_rate": 1e-06,
"loss": 0.2048,
"step": 2748
},
{
"epoch": 1.051242829827916,
"grad_norm": 3.955302607441138,
"learning_rate": 1e-06,
"loss": 0.2415,
"step": 2749
},
{
"epoch": 1.0516252390057361,
"grad_norm": 2.758995515643434,
"learning_rate": 1e-06,
"loss": 0.1459,
"step": 2750
},
{
"epoch": 1.0520076481835563,
"grad_norm": 4.547328441334444,
"learning_rate": 1e-06,
"loss": 0.2896,
"step": 2751
},
{
"epoch": 1.0523900573613767,
"grad_norm": 2.915663864678086,
"learning_rate": 1e-06,
"loss": 0.0652,
"step": 2752
},
{
"epoch": 1.0527724665391969,
"grad_norm": 5.872634736225061,
"learning_rate": 1e-06,
"loss": 0.1742,
"step": 2753
},
{
"epoch": 1.0531548757170173,
"grad_norm": 4.331522905743822,
"learning_rate": 1e-06,
"loss": 0.5289,
"step": 2754
},
{
"epoch": 1.0535372848948374,
"grad_norm": 3.6470304140580527,
"learning_rate": 1e-06,
"loss": 0.5093,
"step": 2755
},
{
"epoch": 1.0539196940726578,
"grad_norm": 2.853255597538155,
"learning_rate": 1e-06,
"loss": 0.3768,
"step": 2756
},
{
"epoch": 1.054302103250478,
"grad_norm": 2.253852407251978,
"learning_rate": 1e-06,
"loss": 0.3574,
"step": 2757
},
{
"epoch": 1.0546845124282982,
"grad_norm": 2.6423868475212187,
"learning_rate": 1e-06,
"loss": 0.2847,
"step": 2758
},
{
"epoch": 1.0550669216061186,
"grad_norm": 3.174076162182849,
"learning_rate": 1e-06,
"loss": 0.4799,
"step": 2759
},
{
"epoch": 1.0554493307839388,
"grad_norm": 2.041361831432893,
"learning_rate": 1e-06,
"loss": 0.1196,
"step": 2760
},
{
"epoch": 1.0558317399617592,
"grad_norm": 2.6674945460141313,
"learning_rate": 1e-06,
"loss": 0.155,
"step": 2761
},
{
"epoch": 1.0562141491395793,
"grad_norm": 2.9281701799976285,
"learning_rate": 1e-06,
"loss": 0.1407,
"step": 2762
},
{
"epoch": 1.0565965583173995,
"grad_norm": 3.137617198685308,
"learning_rate": 1e-06,
"loss": 0.1237,
"step": 2763
},
{
"epoch": 1.05697896749522,
"grad_norm": 2.9902279964007734,
"learning_rate": 1e-06,
"loss": 0.0829,
"step": 2764
},
{
"epoch": 1.05736137667304,
"grad_norm": 3.7043525817923886,
"learning_rate": 1e-06,
"loss": 0.1382,
"step": 2765
},
{
"epoch": 1.0577437858508605,
"grad_norm": 6.382018769497528,
"learning_rate": 1e-06,
"loss": 0.5412,
"step": 2766
},
{
"epoch": 1.0581261950286807,
"grad_norm": 3.774891701464836,
"learning_rate": 1e-06,
"loss": 0.4228,
"step": 2767
},
{
"epoch": 1.058508604206501,
"grad_norm": 3.79119988946345,
"learning_rate": 1e-06,
"loss": 0.4332,
"step": 2768
},
{
"epoch": 1.0588910133843212,
"grad_norm": 2.4406959904348207,
"learning_rate": 1e-06,
"loss": 0.2894,
"step": 2769
},
{
"epoch": 1.0592734225621414,
"grad_norm": 3.5431391329104103,
"learning_rate": 1e-06,
"loss": 0.3917,
"step": 2770
},
{
"epoch": 1.0596558317399618,
"grad_norm": 3.263546222708345,
"learning_rate": 1e-06,
"loss": 0.2839,
"step": 2771
},
{
"epoch": 1.060038240917782,
"grad_norm": 3.3798940989590522,
"learning_rate": 1e-06,
"loss": 0.2712,
"step": 2772
},
{
"epoch": 1.0604206500956024,
"grad_norm": 2.3786540782549745,
"learning_rate": 1e-06,
"loss": 0.1636,
"step": 2773
},
{
"epoch": 1.0608030592734226,
"grad_norm": 3.8724861604801575,
"learning_rate": 1e-06,
"loss": 0.268,
"step": 2774
},
{
"epoch": 1.0611854684512427,
"grad_norm": 3.4577965530462174,
"learning_rate": 1e-06,
"loss": 0.1355,
"step": 2775
},
{
"epoch": 1.0615678776290631,
"grad_norm": 2.6112711521458847,
"learning_rate": 1e-06,
"loss": 0.0885,
"step": 2776
},
{
"epoch": 1.0619502868068833,
"grad_norm": 3.3136666600623528,
"learning_rate": 1e-06,
"loss": 0.0643,
"step": 2777
},
{
"epoch": 1.0623326959847037,
"grad_norm": 4.084531699037168,
"learning_rate": 1e-06,
"loss": 0.1348,
"step": 2778
},
{
"epoch": 1.0627151051625239,
"grad_norm": 3.2009461970957367,
"learning_rate": 1e-06,
"loss": 0.2651,
"step": 2779
},
{
"epoch": 1.063097514340344,
"grad_norm": 3.2590256588801596,
"learning_rate": 1e-06,
"loss": 0.386,
"step": 2780
},
{
"epoch": 1.0634799235181644,
"grad_norm": 2.7223852999707043,
"learning_rate": 1e-06,
"loss": 0.463,
"step": 2781
},
{
"epoch": 1.0638623326959846,
"grad_norm": 2.9683073014908654,
"learning_rate": 1e-06,
"loss": 0.2778,
"step": 2782
},
{
"epoch": 1.064244741873805,
"grad_norm": 2.1593894770312243,
"learning_rate": 1e-06,
"loss": 0.2164,
"step": 2783
},
{
"epoch": 1.0646271510516252,
"grad_norm": 4.2599906179835205,
"learning_rate": 1e-06,
"loss": 0.3968,
"step": 2784
},
{
"epoch": 1.0650095602294456,
"grad_norm": 3.261442581637437,
"learning_rate": 1e-06,
"loss": 0.2604,
"step": 2785
},
{
"epoch": 1.0653919694072658,
"grad_norm": 2.456470419430958,
"learning_rate": 1e-06,
"loss": 0.2151,
"step": 2786
},
{
"epoch": 1.065774378585086,
"grad_norm": 5.335437041867739,
"learning_rate": 1e-06,
"loss": 0.2321,
"step": 2787
},
{
"epoch": 1.0661567877629063,
"grad_norm": 2.525007485860171,
"learning_rate": 1e-06,
"loss": 0.0643,
"step": 2788
},
{
"epoch": 1.0665391969407265,
"grad_norm": 3.073735239087343,
"learning_rate": 1e-06,
"loss": 0.0812,
"step": 2789
},
{
"epoch": 1.066921606118547,
"grad_norm": 3.9811808629522973,
"learning_rate": 1e-06,
"loss": 0.0931,
"step": 2790
},
{
"epoch": 1.067304015296367,
"grad_norm": 7.415904259517387,
"learning_rate": 1e-06,
"loss": 0.5791,
"step": 2791
},
{
"epoch": 1.0676864244741875,
"grad_norm": 3.223305006905413,
"learning_rate": 1e-06,
"loss": 0.3976,
"step": 2792
},
{
"epoch": 1.0680688336520077,
"grad_norm": 3.407348455696182,
"learning_rate": 1e-06,
"loss": 0.4962,
"step": 2793
},
{
"epoch": 1.0684512428298278,
"grad_norm": 2.289743784268278,
"learning_rate": 1e-06,
"loss": 0.1987,
"step": 2794
},
{
"epoch": 1.0688336520076482,
"grad_norm": 3.8234927430271717,
"learning_rate": 1e-06,
"loss": 0.3451,
"step": 2795
},
{
"epoch": 1.0692160611854684,
"grad_norm": 3.451374559652021,
"learning_rate": 1e-06,
"loss": 0.3375,
"step": 2796
},
{
"epoch": 1.0695984703632888,
"grad_norm": 1.776191354598207,
"learning_rate": 1e-06,
"loss": 0.3698,
"step": 2797
},
{
"epoch": 1.069980879541109,
"grad_norm": 2.7212399010039374,
"learning_rate": 1e-06,
"loss": 0.2403,
"step": 2798
},
{
"epoch": 1.0703632887189292,
"grad_norm": 3.5523219023805153,
"learning_rate": 1e-06,
"loss": 0.1581,
"step": 2799
},
{
"epoch": 1.0707456978967496,
"grad_norm": 3.3041342414204586,
"learning_rate": 1e-06,
"loss": 0.1054,
"step": 2800
},
{
"epoch": 1.0711281070745697,
"grad_norm": 3.322390778800794,
"learning_rate": 1e-06,
"loss": 0.1228,
"step": 2801
},
{
"epoch": 1.0715105162523901,
"grad_norm": 3.685726709228734,
"learning_rate": 1e-06,
"loss": 0.0915,
"step": 2802
},
{
"epoch": 1.0718929254302103,
"grad_norm": 3.7334297292616827,
"learning_rate": 1e-06,
"loss": 0.2692,
"step": 2803
},
{
"epoch": 1.0722753346080305,
"grad_norm": 4.571780863879412,
"learning_rate": 1e-06,
"loss": 0.4868,
"step": 2804
},
{
"epoch": 1.0726577437858509,
"grad_norm": 3.4253977579413846,
"learning_rate": 1e-06,
"loss": 0.3587,
"step": 2805
},
{
"epoch": 1.073040152963671,
"grad_norm": 5.506511994622715,
"learning_rate": 1e-06,
"loss": 0.322,
"step": 2806
},
{
"epoch": 1.0734225621414915,
"grad_norm": 1.916366747181914,
"learning_rate": 1e-06,
"loss": 0.2798,
"step": 2807
},
{
"epoch": 1.0738049713193116,
"grad_norm": 2.77367810562294,
"learning_rate": 1e-06,
"loss": 0.3029,
"step": 2808
},
{
"epoch": 1.074187380497132,
"grad_norm": 1.845260841622034,
"learning_rate": 1e-06,
"loss": 0.2007,
"step": 2809
},
{
"epoch": 1.0745697896749522,
"grad_norm": 2.767604485859373,
"learning_rate": 1e-06,
"loss": 0.1358,
"step": 2810
},
{
"epoch": 1.0749521988527724,
"grad_norm": 3.3942540871007716,
"learning_rate": 1e-06,
"loss": 0.1765,
"step": 2811
},
{
"epoch": 1.0753346080305928,
"grad_norm": 4.302997901385853,
"learning_rate": 1e-06,
"loss": 0.4149,
"step": 2812
},
{
"epoch": 1.075717017208413,
"grad_norm": 6.473597382733225,
"learning_rate": 1e-06,
"loss": 0.2666,
"step": 2813
},
{
"epoch": 1.0760994263862333,
"grad_norm": 2.8171827539455827,
"learning_rate": 1e-06,
"loss": 0.0674,
"step": 2814
},
{
"epoch": 1.0764818355640535,
"grad_norm": 5.56599882006727,
"learning_rate": 1e-06,
"loss": 0.1671,
"step": 2815
},
{
"epoch": 1.076864244741874,
"grad_norm": 7.30450915766582,
"learning_rate": 1e-06,
"loss": 0.3735,
"step": 2816
},
{
"epoch": 1.077246653919694,
"grad_norm": 3.423064560731871,
"learning_rate": 1e-06,
"loss": 0.4754,
"step": 2817
},
{
"epoch": 1.0776290630975143,
"grad_norm": 3.2174641531464228,
"learning_rate": 1e-06,
"loss": 0.4532,
"step": 2818
},
{
"epoch": 1.0780114722753347,
"grad_norm": 3.009919139751804,
"learning_rate": 1e-06,
"loss": 0.2953,
"step": 2819
},
{
"epoch": 1.0783938814531548,
"grad_norm": 3.5596707804751624,
"learning_rate": 1e-06,
"loss": 0.4202,
"step": 2820
},
{
"epoch": 1.0787762906309752,
"grad_norm": 3.4648525577678355,
"learning_rate": 1e-06,
"loss": 0.2884,
"step": 2821
},
{
"epoch": 1.0791586998087954,
"grad_norm": 2.6964693544564353,
"learning_rate": 1e-06,
"loss": 0.2034,
"step": 2822
},
{
"epoch": 1.0795411089866156,
"grad_norm": 2.2880146635826097,
"learning_rate": 1e-06,
"loss": 0.1792,
"step": 2823
},
{
"epoch": 1.079923518164436,
"grad_norm": 3.084925232016781,
"learning_rate": 1e-06,
"loss": 0.1477,
"step": 2824
},
{
"epoch": 1.0803059273422562,
"grad_norm": 3.3123235115682896,
"learning_rate": 1e-06,
"loss": 0.1193,
"step": 2825
},
{
"epoch": 1.0806883365200766,
"grad_norm": 3.385762890449034,
"learning_rate": 1e-06,
"loss": 0.0661,
"step": 2826
},
{
"epoch": 1.0810707456978967,
"grad_norm": 3.7251373060893673,
"learning_rate": 1e-06,
"loss": 0.1037,
"step": 2827
},
{
"epoch": 1.081453154875717,
"grad_norm": 5.675358482835371,
"learning_rate": 1e-06,
"loss": 0.2238,
"step": 2828
},
{
"epoch": 1.0818355640535373,
"grad_norm": 5.755246298364127,
"learning_rate": 1e-06,
"loss": 0.3204,
"step": 2829
},
{
"epoch": 1.0822179732313575,
"grad_norm": 4.0569034467401215,
"learning_rate": 1e-06,
"loss": 0.5256,
"step": 2830
},
{
"epoch": 1.0826003824091779,
"grad_norm": 2.688257775928962,
"learning_rate": 1e-06,
"loss": 0.2783,
"step": 2831
},
{
"epoch": 1.082982791586998,
"grad_norm": 2.166301794384953,
"learning_rate": 1e-06,
"loss": 0.2526,
"step": 2832
},
{
"epoch": 1.0833652007648185,
"grad_norm": 3.2912193348488654,
"learning_rate": 1e-06,
"loss": 0.6082,
"step": 2833
},
{
"epoch": 1.0837476099426386,
"grad_norm": 3.5575385094151377,
"learning_rate": 1e-06,
"loss": 0.4073,
"step": 2834
},
{
"epoch": 1.0841300191204588,
"grad_norm": 4.088115979365335,
"learning_rate": 1e-06,
"loss": 0.1996,
"step": 2835
},
{
"epoch": 1.0845124282982792,
"grad_norm": 2.1587060402799323,
"learning_rate": 1e-06,
"loss": 0.0898,
"step": 2836
},
{
"epoch": 1.0848948374760994,
"grad_norm": 3.0045968441273025,
"learning_rate": 1e-06,
"loss": 0.1477,
"step": 2837
},
{
"epoch": 1.0852772466539198,
"grad_norm": 4.259853608745767,
"learning_rate": 1e-06,
"loss": 0.1283,
"step": 2838
},
{
"epoch": 1.08565965583174,
"grad_norm": 4.144053756144276,
"learning_rate": 1e-06,
"loss": 0.1443,
"step": 2839
},
{
"epoch": 1.0860420650095601,
"grad_norm": 4.5304088995766465,
"learning_rate": 1e-06,
"loss": 0.1254,
"step": 2840
},
{
"epoch": 1.0864244741873805,
"grad_norm": 7.951371934660744,
"learning_rate": 1e-06,
"loss": 0.2675,
"step": 2841
},
{
"epoch": 1.0868068833652007,
"grad_norm": 4.102531275098238,
"learning_rate": 1e-06,
"loss": 0.4678,
"step": 2842
},
{
"epoch": 1.087189292543021,
"grad_norm": 2.8548635110113243,
"learning_rate": 1e-06,
"loss": 0.2682,
"step": 2843
},
{
"epoch": 1.0875717017208413,
"grad_norm": 2.550358174918483,
"learning_rate": 1e-06,
"loss": 0.1595,
"step": 2844
},
{
"epoch": 1.0879541108986617,
"grad_norm": 2.885750284199724,
"learning_rate": 1e-06,
"loss": 0.4214,
"step": 2845
},
{
"epoch": 1.0883365200764819,
"grad_norm": 2.413077045271455,
"learning_rate": 1e-06,
"loss": 0.3292,
"step": 2846
},
{
"epoch": 1.088718929254302,
"grad_norm": 2.977832913560461,
"learning_rate": 1e-06,
"loss": 0.2312,
"step": 2847
},
{
"epoch": 1.0891013384321224,
"grad_norm": 3.0670257137030372,
"learning_rate": 1e-06,
"loss": 0.2079,
"step": 2848
},
{
"epoch": 1.0894837476099426,
"grad_norm": 3.567873866935165,
"learning_rate": 1e-06,
"loss": 0.3071,
"step": 2849
},
{
"epoch": 1.089866156787763,
"grad_norm": 3.2884373014022796,
"learning_rate": 1e-06,
"loss": 0.216,
"step": 2850
},
{
"epoch": 1.0902485659655832,
"grad_norm": 3.167269515493883,
"learning_rate": 1e-06,
"loss": 0.0664,
"step": 2851
},
{
"epoch": 1.0906309751434033,
"grad_norm": 4.1067322330277465,
"learning_rate": 1e-06,
"loss": 0.1224,
"step": 2852
},
{
"epoch": 1.0910133843212237,
"grad_norm": 4.817414585815298,
"learning_rate": 1e-06,
"loss": 0.4915,
"step": 2853
},
{
"epoch": 1.091395793499044,
"grad_norm": 3.696390587484518,
"learning_rate": 1e-06,
"loss": 0.4389,
"step": 2854
},
{
"epoch": 1.0917782026768643,
"grad_norm": 3.057344105848634,
"learning_rate": 1e-06,
"loss": 0.4314,
"step": 2855
},
{
"epoch": 1.0921606118546845,
"grad_norm": 3.1848468584405674,
"learning_rate": 1e-06,
"loss": 0.3289,
"step": 2856
},
{
"epoch": 1.0925430210325047,
"grad_norm": 3.114695688387905,
"learning_rate": 1e-06,
"loss": 0.3441,
"step": 2857
},
{
"epoch": 1.092925430210325,
"grad_norm": 6.4186641341931985,
"learning_rate": 1e-06,
"loss": 0.4204,
"step": 2858
},
{
"epoch": 1.0933078393881452,
"grad_norm": 3.436542585730967,
"learning_rate": 1e-06,
"loss": 0.1804,
"step": 2859
},
{
"epoch": 1.0936902485659656,
"grad_norm": 2.5285077723571674,
"learning_rate": 1e-06,
"loss": 0.2002,
"step": 2860
},
{
"epoch": 1.0940726577437858,
"grad_norm": 2.581862820716856,
"learning_rate": 1e-06,
"loss": 0.1119,
"step": 2861
},
{
"epoch": 1.0944550669216062,
"grad_norm": 3.2284916982439156,
"learning_rate": 1e-06,
"loss": 0.1461,
"step": 2862
},
{
"epoch": 1.0948374760994264,
"grad_norm": 3.2432258138317156,
"learning_rate": 1e-06,
"loss": 0.1405,
"step": 2863
},
{
"epoch": 1.0952198852772466,
"grad_norm": 3.056057595903582,
"learning_rate": 1e-06,
"loss": 0.0567,
"step": 2864
},
{
"epoch": 1.095602294455067,
"grad_norm": 3.406085929068903,
"learning_rate": 1e-06,
"loss": 0.0868,
"step": 2865
},
{
"epoch": 1.0959847036328871,
"grad_norm": 5.2675466228896735,
"learning_rate": 1e-06,
"loss": 0.4209,
"step": 2866
},
{
"epoch": 1.0963671128107075,
"grad_norm": 3.495768577754933,
"learning_rate": 1e-06,
"loss": 0.5578,
"step": 2867
},
{
"epoch": 1.0967495219885277,
"grad_norm": 2.3411254429309367,
"learning_rate": 1e-06,
"loss": 0.3557,
"step": 2868
},
{
"epoch": 1.097131931166348,
"grad_norm": 3.0691303957486715,
"learning_rate": 1e-06,
"loss": 0.3543,
"step": 2869
},
{
"epoch": 1.0975143403441683,
"grad_norm": 3.065784639174185,
"learning_rate": 1e-06,
"loss": 0.3856,
"step": 2870
},
{
"epoch": 1.0978967495219885,
"grad_norm": 2.994955908652979,
"learning_rate": 1e-06,
"loss": 0.2518,
"step": 2871
},
{
"epoch": 1.0982791586998089,
"grad_norm": 4.563174472232909,
"learning_rate": 1e-06,
"loss": 0.1964,
"step": 2872
},
{
"epoch": 1.098661567877629,
"grad_norm": 3.4571755244435485,
"learning_rate": 1e-06,
"loss": 0.2262,
"step": 2873
},
{
"epoch": 1.0990439770554494,
"grad_norm": 2.5488111005777228,
"learning_rate": 1e-06,
"loss": 0.1324,
"step": 2874
},
{
"epoch": 1.0994263862332696,
"grad_norm": 3.945085394337583,
"learning_rate": 1e-06,
"loss": 0.1595,
"step": 2875
},
{
"epoch": 1.0998087954110898,
"grad_norm": 2.837069030836222,
"learning_rate": 1e-06,
"loss": 0.099,
"step": 2876
},
{
"epoch": 1.1001912045889102,
"grad_norm": 3.908802632757324,
"learning_rate": 1e-06,
"loss": 0.0899,
"step": 2877
},
{
"epoch": 1.1005736137667304,
"grad_norm": 5.114341818188961,
"learning_rate": 1e-06,
"loss": 0.3208,
"step": 2878
},
{
"epoch": 1.1009560229445507,
"grad_norm": 3.858759904409437,
"learning_rate": 1e-06,
"loss": 0.4088,
"step": 2879
},
{
"epoch": 1.101338432122371,
"grad_norm": 2.6757349163415456,
"learning_rate": 1e-06,
"loss": 0.2739,
"step": 2880
},
{
"epoch": 1.101720841300191,
"grad_norm": 2.6258524464265034,
"learning_rate": 1e-06,
"loss": 0.2561,
"step": 2881
},
{
"epoch": 1.1021032504780115,
"grad_norm": 4.256005700000015,
"learning_rate": 1e-06,
"loss": 0.2905,
"step": 2882
},
{
"epoch": 1.1024856596558317,
"grad_norm": 2.93229359998803,
"learning_rate": 1e-06,
"loss": 0.2834,
"step": 2883
},
{
"epoch": 1.102868068833652,
"grad_norm": 3.475864434613361,
"learning_rate": 1e-06,
"loss": 0.2758,
"step": 2884
},
{
"epoch": 1.1032504780114722,
"grad_norm": 2.295733836580188,
"learning_rate": 1e-06,
"loss": 0.1278,
"step": 2885
},
{
"epoch": 1.1036328871892926,
"grad_norm": 3.369627880300253,
"learning_rate": 1e-06,
"loss": 0.46,
"step": 2886
},
{
"epoch": 1.1040152963671128,
"grad_norm": 2.5749161604778497,
"learning_rate": 1e-06,
"loss": 0.0795,
"step": 2887
},
{
"epoch": 1.104397705544933,
"grad_norm": 5.368106058246976,
"learning_rate": 1e-06,
"loss": 0.3168,
"step": 2888
},
{
"epoch": 1.1047801147227534,
"grad_norm": 3.4822033328439526,
"learning_rate": 1e-06,
"loss": 0.0966,
"step": 2889
},
{
"epoch": 1.1051625239005736,
"grad_norm": 5.389917476496499,
"learning_rate": 1e-06,
"loss": 0.1366,
"step": 2890
},
{
"epoch": 1.105544933078394,
"grad_norm": 7.55366188827401,
"learning_rate": 1e-06,
"loss": 0.3474,
"step": 2891
},
{
"epoch": 1.1059273422562141,
"grad_norm": 4.755177135998886,
"learning_rate": 1e-06,
"loss": 0.627,
"step": 2892
},
{
"epoch": 1.1063097514340343,
"grad_norm": 2.0768261055438555,
"learning_rate": 1e-06,
"loss": 0.3692,
"step": 2893
},
{
"epoch": 1.1066921606118547,
"grad_norm": 3.225032268037455,
"learning_rate": 1e-06,
"loss": 0.3745,
"step": 2894
},
{
"epoch": 1.107074569789675,
"grad_norm": 3.9894524030723058,
"learning_rate": 1e-06,
"loss": 0.4802,
"step": 2895
},
{
"epoch": 1.1074569789674953,
"grad_norm": 3.2782512222295908,
"learning_rate": 1e-06,
"loss": 0.2156,
"step": 2896
},
{
"epoch": 1.1078393881453155,
"grad_norm": 3.825465185264008,
"learning_rate": 1e-06,
"loss": 0.2157,
"step": 2897
},
{
"epoch": 1.1082217973231359,
"grad_norm": 3.158481526651568,
"learning_rate": 1e-06,
"loss": 0.3175,
"step": 2898
},
{
"epoch": 1.108604206500956,
"grad_norm": 2.5433234979615165,
"learning_rate": 1e-06,
"loss": 0.0968,
"step": 2899
},
{
"epoch": 1.1089866156787762,
"grad_norm": 2.9339906548453296,
"learning_rate": 1e-06,
"loss": 0.0736,
"step": 2900
},
{
"epoch": 1.1093690248565966,
"grad_norm": 3.7096563115261763,
"learning_rate": 1e-06,
"loss": 0.1069,
"step": 2901
},
{
"epoch": 1.1097514340344168,
"grad_norm": 4.340461638900813,
"learning_rate": 1e-06,
"loss": 0.0848,
"step": 2902
},
{
"epoch": 1.1101338432122372,
"grad_norm": 5.716557984204094,
"learning_rate": 1e-06,
"loss": 0.2638,
"step": 2903
},
{
"epoch": 1.1105162523900574,
"grad_norm": 4.055877922402935,
"learning_rate": 1e-06,
"loss": 0.6603,
"step": 2904
},
{
"epoch": 1.1108986615678775,
"grad_norm": 2.792949030879908,
"learning_rate": 1e-06,
"loss": 0.3407,
"step": 2905
},
{
"epoch": 1.111281070745698,
"grad_norm": 3.4354327835127525,
"learning_rate": 1e-06,
"loss": 0.3378,
"step": 2906
},
{
"epoch": 1.111663479923518,
"grad_norm": 3.996718968379817,
"learning_rate": 1e-06,
"loss": 0.3919,
"step": 2907
},
{
"epoch": 1.1120458891013385,
"grad_norm": 2.982178363963938,
"learning_rate": 1e-06,
"loss": 0.3849,
"step": 2908
},
{
"epoch": 1.1124282982791587,
"grad_norm": 3.227012030296255,
"learning_rate": 1e-06,
"loss": 0.1708,
"step": 2909
},
{
"epoch": 1.1128107074569789,
"grad_norm": 2.6302030859161505,
"learning_rate": 1e-06,
"loss": 0.31,
"step": 2910
},
{
"epoch": 1.1131931166347993,
"grad_norm": 3.635419144215145,
"learning_rate": 1e-06,
"loss": 0.2621,
"step": 2911
},
{
"epoch": 1.1135755258126194,
"grad_norm": 2.9696596057171476,
"learning_rate": 1e-06,
"loss": 0.1988,
"step": 2912
},
{
"epoch": 1.1139579349904398,
"grad_norm": 4.434582248008877,
"learning_rate": 1e-06,
"loss": 0.1347,
"step": 2913
},
{
"epoch": 1.11434034416826,
"grad_norm": 3.670918181727296,
"learning_rate": 1e-06,
"loss": 0.082,
"step": 2914
},
{
"epoch": 1.1147227533460804,
"grad_norm": 4.325960685449046,
"learning_rate": 1e-06,
"loss": 0.1222,
"step": 2915
},
{
"epoch": 1.1151051625239006,
"grad_norm": 7.413527377948621,
"learning_rate": 1e-06,
"loss": 0.2725,
"step": 2916
},
{
"epoch": 1.1154875717017207,
"grad_norm": 4.250834214859102,
"learning_rate": 1e-06,
"loss": 0.784,
"step": 2917
},
{
"epoch": 1.1158699808795411,
"grad_norm": 3.401164607133041,
"learning_rate": 1e-06,
"loss": 0.4209,
"step": 2918
},
{
"epoch": 1.1162523900573613,
"grad_norm": 2.6646607921358387,
"learning_rate": 1e-06,
"loss": 0.4007,
"step": 2919
},
{
"epoch": 1.1166347992351817,
"grad_norm": 2.868380015331566,
"learning_rate": 1e-06,
"loss": 0.4022,
"step": 2920
},
{
"epoch": 1.117017208413002,
"grad_norm": 2.8090205392219025,
"learning_rate": 1e-06,
"loss": 0.2514,
"step": 2921
},
{
"epoch": 1.1173996175908223,
"grad_norm": 3.8349616835369784,
"learning_rate": 1e-06,
"loss": 0.3427,
"step": 2922
},
{
"epoch": 1.1177820267686425,
"grad_norm": 2.74083900442568,
"learning_rate": 1e-06,
"loss": 0.2086,
"step": 2923
},
{
"epoch": 1.1181644359464626,
"grad_norm": 2.8517920832772288,
"learning_rate": 1e-06,
"loss": 0.2163,
"step": 2924
},
{
"epoch": 1.118546845124283,
"grad_norm": 3.2410084982162966,
"learning_rate": 1e-06,
"loss": 0.2256,
"step": 2925
},
{
"epoch": 1.1189292543021032,
"grad_norm": 2.199672401485876,
"learning_rate": 1e-06,
"loss": 0.0648,
"step": 2926
},
{
"epoch": 1.1193116634799236,
"grad_norm": 2.89625339525968,
"learning_rate": 1e-06,
"loss": 0.0964,
"step": 2927
},
{
"epoch": 1.1196940726577438,
"grad_norm": 3.675327187849286,
"learning_rate": 1e-06,
"loss": 0.1628,
"step": 2928
},
{
"epoch": 1.120076481835564,
"grad_norm": 5.544467801282905,
"learning_rate": 1e-06,
"loss": 0.454,
"step": 2929
},
{
"epoch": 1.1204588910133844,
"grad_norm": 4.969388332989557,
"learning_rate": 1e-06,
"loss": 0.486,
"step": 2930
},
{
"epoch": 1.1208413001912045,
"grad_norm": 2.787810485380719,
"learning_rate": 1e-06,
"loss": 0.1456,
"step": 2931
},
{
"epoch": 1.121223709369025,
"grad_norm": 3.13911318094859,
"learning_rate": 1e-06,
"loss": 0.2817,
"step": 2932
},
{
"epoch": 1.121606118546845,
"grad_norm": 3.0418354318560086,
"learning_rate": 1e-06,
"loss": 0.2556,
"step": 2933
},
{
"epoch": 1.1219885277246653,
"grad_norm": 3.7041959220037035,
"learning_rate": 1e-06,
"loss": 0.2334,
"step": 2934
},
{
"epoch": 1.1223709369024857,
"grad_norm": 2.518353611350916,
"learning_rate": 1e-06,
"loss": 0.1713,
"step": 2935
},
{
"epoch": 1.1227533460803059,
"grad_norm": 2.864493702728919,
"learning_rate": 1e-06,
"loss": 0.2985,
"step": 2936
},
{
"epoch": 1.1231357552581263,
"grad_norm": 3.3934064434796807,
"learning_rate": 1e-06,
"loss": 0.0911,
"step": 2937
},
{
"epoch": 1.1235181644359464,
"grad_norm": 3.192170853461705,
"learning_rate": 1e-06,
"loss": 0.079,
"step": 2938
},
{
"epoch": 1.1239005736137668,
"grad_norm": 3.362388501215389,
"learning_rate": 1e-06,
"loss": 0.076,
"step": 2939
},
{
"epoch": 1.124282982791587,
"grad_norm": 5.310276957615031,
"learning_rate": 1e-06,
"loss": 0.1389,
"step": 2940
},
{
"epoch": 1.1246653919694072,
"grad_norm": 6.827063995262543,
"learning_rate": 1e-06,
"loss": 0.3134,
"step": 2941
},
{
"epoch": 1.1250478011472276,
"grad_norm": 3.67451674209873,
"learning_rate": 1e-06,
"loss": 0.4647,
"step": 2942
},
{
"epoch": 1.1254302103250478,
"grad_norm": 2.333746862097798,
"learning_rate": 1e-06,
"loss": 0.2564,
"step": 2943
},
{
"epoch": 1.1258126195028682,
"grad_norm": 3.0286945700995,
"learning_rate": 1e-06,
"loss": 0.5127,
"step": 2944
},
{
"epoch": 1.1261950286806883,
"grad_norm": 3.895260052991601,
"learning_rate": 1e-06,
"loss": 0.3841,
"step": 2945
},
{
"epoch": 1.1265774378585087,
"grad_norm": 4.60890159115522,
"learning_rate": 1e-06,
"loss": 0.2758,
"step": 2946
},
{
"epoch": 1.126959847036329,
"grad_norm": 2.299857558110427,
"learning_rate": 1e-06,
"loss": 0.1026,
"step": 2947
},
{
"epoch": 1.127342256214149,
"grad_norm": 2.301337752857731,
"learning_rate": 1e-06,
"loss": 0.0975,
"step": 2948
},
{
"epoch": 1.1277246653919695,
"grad_norm": 2.434746580938376,
"learning_rate": 1e-06,
"loss": 0.1672,
"step": 2949
},
{
"epoch": 1.1281070745697896,
"grad_norm": 3.3248961640400063,
"learning_rate": 1e-06,
"loss": 0.2036,
"step": 2950
},
{
"epoch": 1.12848948374761,
"grad_norm": 2.399819836212055,
"learning_rate": 1e-06,
"loss": 0.0503,
"step": 2951
},
{
"epoch": 1.1288718929254302,
"grad_norm": 8.43861980777982,
"learning_rate": 1e-06,
"loss": 0.1625,
"step": 2952
},
{
"epoch": 1.1292543021032504,
"grad_norm": 5.674700745073927,
"learning_rate": 1e-06,
"loss": 0.1837,
"step": 2953
},
{
"epoch": 1.1296367112810708,
"grad_norm": 2.8559599948293095,
"learning_rate": 1e-06,
"loss": 0.4173,
"step": 2954
},
{
"epoch": 1.130019120458891,
"grad_norm": 3.1598565675077217,
"learning_rate": 1e-06,
"loss": 0.3471,
"step": 2955
},
{
"epoch": 1.1304015296367114,
"grad_norm": 3.5022016140314807,
"learning_rate": 1e-06,
"loss": 0.2841,
"step": 2956
},
{
"epoch": 1.1307839388145315,
"grad_norm": 3.0422650798041744,
"learning_rate": 1e-06,
"loss": 0.3529,
"step": 2957
},
{
"epoch": 1.1311663479923517,
"grad_norm": 3.743453096516027,
"learning_rate": 1e-06,
"loss": 0.3814,
"step": 2958
},
{
"epoch": 1.1315487571701721,
"grad_norm": 3.170667446641155,
"learning_rate": 1e-06,
"loss": 0.3059,
"step": 2959
},
{
"epoch": 1.1319311663479923,
"grad_norm": 3.8093364910779806,
"learning_rate": 1e-06,
"loss": 0.1791,
"step": 2960
},
{
"epoch": 1.1323135755258127,
"grad_norm": 2.7752219034086587,
"learning_rate": 1e-06,
"loss": 0.2031,
"step": 2961
},
{
"epoch": 1.1326959847036329,
"grad_norm": 4.268290040018679,
"learning_rate": 1e-06,
"loss": 0.1393,
"step": 2962
},
{
"epoch": 1.133078393881453,
"grad_norm": 5.73511469173405,
"learning_rate": 1e-06,
"loss": 0.1168,
"step": 2963
},
{
"epoch": 1.1334608030592734,
"grad_norm": 5.463547853576047,
"learning_rate": 1e-06,
"loss": 0.1538,
"step": 2964
},
{
"epoch": 1.1338432122370936,
"grad_norm": 5.990980999178643,
"learning_rate": 1e-06,
"loss": 0.1144,
"step": 2965
},
{
"epoch": 1.134225621414914,
"grad_norm": 4.907622127247335,
"learning_rate": 1e-06,
"loss": 0.2499,
"step": 2966
},
{
"epoch": 1.1346080305927342,
"grad_norm": 2.6854682161775996,
"learning_rate": 1e-06,
"loss": 0.5883,
"step": 2967
},
{
"epoch": 1.1349904397705546,
"grad_norm": 3.890548644025143,
"learning_rate": 1e-06,
"loss": 0.4597,
"step": 2968
},
{
"epoch": 1.1353728489483748,
"grad_norm": 1.8231637260500944,
"learning_rate": 1e-06,
"loss": 0.3064,
"step": 2969
},
{
"epoch": 1.135755258126195,
"grad_norm": 3.5759270593086896,
"learning_rate": 1e-06,
"loss": 0.2339,
"step": 2970
},
{
"epoch": 1.1361376673040153,
"grad_norm": 3.260923807273525,
"learning_rate": 1e-06,
"loss": 0.3713,
"step": 2971
},
{
"epoch": 1.1365200764818355,
"grad_norm": 2.9478360402075094,
"learning_rate": 1e-06,
"loss": 0.1954,
"step": 2972
},
{
"epoch": 1.136902485659656,
"grad_norm": 2.9678518442464363,
"learning_rate": 1e-06,
"loss": 0.3386,
"step": 2973
},
{
"epoch": 1.137284894837476,
"grad_norm": 3.5506315126310657,
"learning_rate": 1e-06,
"loss": 0.1616,
"step": 2974
},
{
"epoch": 1.1376673040152965,
"grad_norm": 2.289140615334694,
"learning_rate": 1e-06,
"loss": 0.1095,
"step": 2975
},
{
"epoch": 1.1380497131931167,
"grad_norm": 3.952317830181921,
"learning_rate": 1e-06,
"loss": 0.25,
"step": 2976
},
{
"epoch": 1.1384321223709368,
"grad_norm": 3.027342174898784,
"learning_rate": 1e-06,
"loss": 0.0998,
"step": 2977
},
{
"epoch": 1.1388145315487572,
"grad_norm": 4.896933876954402,
"learning_rate": 1e-06,
"loss": 0.2392,
"step": 2978
},
{
"epoch": 1.1391969407265774,
"grad_norm": 4.925390144812214,
"learning_rate": 1e-06,
"loss": 0.6317,
"step": 2979
},
{
"epoch": 1.1395793499043978,
"grad_norm": 3.719461196723674,
"learning_rate": 1e-06,
"loss": 0.4982,
"step": 2980
},
{
"epoch": 1.139961759082218,
"grad_norm": 5.013784861285325,
"learning_rate": 1e-06,
"loss": 0.2064,
"step": 2981
},
{
"epoch": 1.1403441682600381,
"grad_norm": 4.531964686364737,
"learning_rate": 1e-06,
"loss": 0.27,
"step": 2982
},
{
"epoch": 1.1407265774378585,
"grad_norm": 2.65648317996501,
"learning_rate": 1e-06,
"loss": 0.3822,
"step": 2983
},
{
"epoch": 1.1411089866156787,
"grad_norm": 2.444294678068928,
"learning_rate": 1e-06,
"loss": 0.3012,
"step": 2984
},
{
"epoch": 1.1414913957934991,
"grad_norm": 3.6912378020428958,
"learning_rate": 1e-06,
"loss": 0.1829,
"step": 2985
},
{
"epoch": 1.1418738049713193,
"grad_norm": 3.5530446710288026,
"learning_rate": 1e-06,
"loss": 0.2182,
"step": 2986
},
{
"epoch": 1.1422562141491395,
"grad_norm": 3.4648946695983605,
"learning_rate": 1e-06,
"loss": 0.1276,
"step": 2987
},
{
"epoch": 1.1426386233269599,
"grad_norm": 3.051578275968885,
"learning_rate": 1e-06,
"loss": 0.1571,
"step": 2988
},
{
"epoch": 1.14302103250478,
"grad_norm": 3.6946144654454556,
"learning_rate": 1e-06,
"loss": 0.0636,
"step": 2989
},
{
"epoch": 1.1434034416826004,
"grad_norm": 3.6885851540309598,
"learning_rate": 1e-06,
"loss": 0.1163,
"step": 2990
},
{
"epoch": 1.1437858508604206,
"grad_norm": 8.472431579115465,
"learning_rate": 1e-06,
"loss": 0.4143,
"step": 2991
},
{
"epoch": 1.144168260038241,
"grad_norm": 4.0249433051434576,
"learning_rate": 1e-06,
"loss": 0.3931,
"step": 2992
},
{
"epoch": 1.1445506692160612,
"grad_norm": 5.038420691019398,
"learning_rate": 1e-06,
"loss": 0.3297,
"step": 2993
},
{
"epoch": 1.1449330783938814,
"grad_norm": 2.810695662811567,
"learning_rate": 1e-06,
"loss": 0.2456,
"step": 2994
},
{
"epoch": 1.1453154875717018,
"grad_norm": 2.493642256854354,
"learning_rate": 1e-06,
"loss": 0.2771,
"step": 2995
},
{
"epoch": 1.145697896749522,
"grad_norm": 3.446172697675935,
"learning_rate": 1e-06,
"loss": 0.2694,
"step": 2996
},
{
"epoch": 1.1460803059273423,
"grad_norm": 3.254186867729784,
"learning_rate": 1e-06,
"loss": 0.2904,
"step": 2997
},
{
"epoch": 1.1464627151051625,
"grad_norm": 3.111249349187137,
"learning_rate": 1e-06,
"loss": 0.27,
"step": 2998
},
{
"epoch": 1.146845124282983,
"grad_norm": 3.145490076061839,
"learning_rate": 1e-06,
"loss": 0.1915,
"step": 2999
},
{
"epoch": 1.147227533460803,
"grad_norm": 3.6650658638774907,
"learning_rate": 1e-06,
"loss": 0.1846,
"step": 3000
},
{
"epoch": 1.147227533460803,
"eval_runtime": 847.684,
"eval_samples_per_second": 1.81,
"eval_steps_per_second": 0.453,
"step": 3000
},
{
"epoch": 1.1476099426386233,
"grad_norm": 5.917576822830572,
"learning_rate": 1e-06,
"loss": 0.1033,
"step": 3001
},
{
"epoch": 1.1479923518164437,
"grad_norm": 4.405933774598463,
"learning_rate": 1e-06,
"loss": 0.0918,
"step": 3002
},
{
"epoch": 1.1483747609942638,
"grad_norm": 7.300066688964566,
"learning_rate": 1e-06,
"loss": 0.2163,
"step": 3003
},
{
"epoch": 1.1487571701720842,
"grad_norm": 3.231642596302886,
"learning_rate": 1e-06,
"loss": 0.4344,
"step": 3004
},
{
"epoch": 1.1491395793499044,
"grad_norm": 2.694351511485435,
"learning_rate": 1e-06,
"loss": 0.2757,
"step": 3005
},
{
"epoch": 1.1495219885277246,
"grad_norm": 2.3460577221116568,
"learning_rate": 1e-06,
"loss": 0.3138,
"step": 3006
},
{
"epoch": 1.149904397705545,
"grad_norm": 2.1963152241452626,
"learning_rate": 1e-06,
"loss": 0.1383,
"step": 3007
},
{
"epoch": 1.1502868068833652,
"grad_norm": 2.6964874802490995,
"learning_rate": 1e-06,
"loss": 0.1382,
"step": 3008
},
{
"epoch": 1.1506692160611856,
"grad_norm": 3.3874123818865773,
"learning_rate": 1e-06,
"loss": 0.2823,
"step": 3009
},
{
"epoch": 1.1510516252390057,
"grad_norm": 3.1424370088538383,
"learning_rate": 1e-06,
"loss": 0.1919,
"step": 3010
},
{
"epoch": 1.151434034416826,
"grad_norm": 3.4296154805401073,
"learning_rate": 1e-06,
"loss": 0.3296,
"step": 3011
},
{
"epoch": 1.1518164435946463,
"grad_norm": 2.6362655137077238,
"learning_rate": 1e-06,
"loss": 0.1152,
"step": 3012
},
{
"epoch": 1.1521988527724665,
"grad_norm": 2.797309202449541,
"learning_rate": 1e-06,
"loss": 0.0643,
"step": 3013
},
{
"epoch": 1.1525812619502869,
"grad_norm": 4.26345490529596,
"learning_rate": 1e-06,
"loss": 0.1267,
"step": 3014
},
{
"epoch": 1.152963671128107,
"grad_norm": 4.9750487512088055,
"learning_rate": 1e-06,
"loss": 0.0788,
"step": 3015
},
{
"epoch": 1.1533460803059272,
"grad_norm": 7.489272265458247,
"learning_rate": 1e-06,
"loss": 0.6298,
"step": 3016
},
{
"epoch": 1.1537284894837476,
"grad_norm": 3.8932575679836057,
"learning_rate": 1e-06,
"loss": 0.397,
"step": 3017
},
{
"epoch": 1.1541108986615678,
"grad_norm": 2.6774029571870215,
"learning_rate": 1e-06,
"loss": 0.2164,
"step": 3018
},
{
"epoch": 1.1544933078393882,
"grad_norm": 2.303157803057292,
"learning_rate": 1e-06,
"loss": 0.1739,
"step": 3019
},
{
"epoch": 1.1548757170172084,
"grad_norm": 3.7227835698485845,
"learning_rate": 1e-06,
"loss": 0.3657,
"step": 3020
},
{
"epoch": 1.1552581261950288,
"grad_norm": 3.15923417529325,
"learning_rate": 1e-06,
"loss": 0.3346,
"step": 3021
},
{
"epoch": 1.155640535372849,
"grad_norm": 2.8229112988710687,
"learning_rate": 1e-06,
"loss": 0.1501,
"step": 3022
},
{
"epoch": 1.1560229445506693,
"grad_norm": 2.5378042066908555,
"learning_rate": 1e-06,
"loss": 0.2735,
"step": 3023
},
{
"epoch": 1.1564053537284895,
"grad_norm": 3.1087593734348347,
"learning_rate": 1e-06,
"loss": 0.1523,
"step": 3024
},
{
"epoch": 1.1567877629063097,
"grad_norm": 3.8077057257996065,
"learning_rate": 1e-06,
"loss": 0.1377,
"step": 3025
},
{
"epoch": 1.15717017208413,
"grad_norm": 3.5215631815745247,
"learning_rate": 1e-06,
"loss": 0.1032,
"step": 3026
},
{
"epoch": 1.1575525812619503,
"grad_norm": 3.840747311710906,
"learning_rate": 1e-06,
"loss": 0.0706,
"step": 3027
},
{
"epoch": 1.1579349904397707,
"grad_norm": 6.1228736767769085,
"learning_rate": 1e-06,
"loss": 0.1465,
"step": 3028
},
{
"epoch": 1.1583173996175908,
"grad_norm": 3.953383365676891,
"learning_rate": 1e-06,
"loss": 0.5059,
"step": 3029
},
{
"epoch": 1.158699808795411,
"grad_norm": 3.1208554058711964,
"learning_rate": 1e-06,
"loss": 0.4097,
"step": 3030
},
{
"epoch": 1.1590822179732314,
"grad_norm": 3.3829022895578587,
"learning_rate": 1e-06,
"loss": 0.453,
"step": 3031
},
{
"epoch": 1.1594646271510516,
"grad_norm": 2.4288076297914083,
"learning_rate": 1e-06,
"loss": 0.2809,
"step": 3032
},
{
"epoch": 1.159847036328872,
"grad_norm": 2.6608344514507745,
"learning_rate": 1e-06,
"loss": 0.2809,
"step": 3033
},
{
"epoch": 1.1602294455066922,
"grad_norm": 4.02149955224188,
"learning_rate": 1e-06,
"loss": 0.2122,
"step": 3034
},
{
"epoch": 1.1606118546845123,
"grad_norm": 3.1588094172986207,
"learning_rate": 1e-06,
"loss": 0.2068,
"step": 3035
},
{
"epoch": 1.1609942638623327,
"grad_norm": 3.170947611115864,
"learning_rate": 1e-06,
"loss": 0.1697,
"step": 3036
},
{
"epoch": 1.161376673040153,
"grad_norm": 2.618345818476673,
"learning_rate": 1e-06,
"loss": 0.2284,
"step": 3037
},
{
"epoch": 1.1617590822179733,
"grad_norm": 3.863046482771306,
"learning_rate": 1e-06,
"loss": 0.1383,
"step": 3038
},
{
"epoch": 1.1621414913957935,
"grad_norm": 4.332092009549107,
"learning_rate": 1e-06,
"loss": 0.2665,
"step": 3039
},
{
"epoch": 1.1625239005736137,
"grad_norm": 3.7784688510718873,
"learning_rate": 1e-06,
"loss": 0.1089,
"step": 3040
},
{
"epoch": 1.162906309751434,
"grad_norm": 4.3595318509668814,
"learning_rate": 1e-06,
"loss": 0.3529,
"step": 3041
},
{
"epoch": 1.1632887189292542,
"grad_norm": 3.421964182627793,
"learning_rate": 1e-06,
"loss": 0.3063,
"step": 3042
},
{
"epoch": 1.1636711281070746,
"grad_norm": 2.8840746093919036,
"learning_rate": 1e-06,
"loss": 0.3455,
"step": 3043
},
{
"epoch": 1.1640535372848948,
"grad_norm": 3.5628827876517244,
"learning_rate": 1e-06,
"loss": 0.3037,
"step": 3044
},
{
"epoch": 1.1644359464627152,
"grad_norm": 4.2031746900379465,
"learning_rate": 1e-06,
"loss": 0.3853,
"step": 3045
},
{
"epoch": 1.1648183556405354,
"grad_norm": 3.292110670702286,
"learning_rate": 1e-06,
"loss": 0.1813,
"step": 3046
},
{
"epoch": 1.1652007648183555,
"grad_norm": 2.308336582479933,
"learning_rate": 1e-06,
"loss": 0.1946,
"step": 3047
},
{
"epoch": 1.165583173996176,
"grad_norm": 2.523929512551048,
"learning_rate": 1e-06,
"loss": 0.187,
"step": 3048
},
{
"epoch": 1.1659655831739961,
"grad_norm": 3.2102688652406965,
"learning_rate": 1e-06,
"loss": 0.1661,
"step": 3049
},
{
"epoch": 1.1663479923518165,
"grad_norm": 3.2124555948357485,
"learning_rate": 1e-06,
"loss": 0.1656,
"step": 3050
},
{
"epoch": 1.1667304015296367,
"grad_norm": 3.8921057399644803,
"learning_rate": 1e-06,
"loss": 0.1297,
"step": 3051
},
{
"epoch": 1.167112810707457,
"grad_norm": 3.6307160922256356,
"learning_rate": 1e-06,
"loss": 0.1017,
"step": 3052
},
{
"epoch": 1.1674952198852773,
"grad_norm": 7.806813360063579,
"learning_rate": 1e-06,
"loss": 0.2513,
"step": 3053
},
{
"epoch": 1.1678776290630974,
"grad_norm": 4.472717444655059,
"learning_rate": 1e-06,
"loss": 0.4653,
"step": 3054
},
{
"epoch": 1.1682600382409178,
"grad_norm": 3.798022438260248,
"learning_rate": 1e-06,
"loss": 0.4606,
"step": 3055
},
{
"epoch": 1.168642447418738,
"grad_norm": 3.1168322826859893,
"learning_rate": 1e-06,
"loss": 0.4191,
"step": 3056
},
{
"epoch": 1.1690248565965584,
"grad_norm": 2.7548288085144588,
"learning_rate": 1e-06,
"loss": 0.2374,
"step": 3057
},
{
"epoch": 1.1694072657743786,
"grad_norm": 3.4893264144529557,
"learning_rate": 1e-06,
"loss": 0.3185,
"step": 3058
},
{
"epoch": 1.1697896749521988,
"grad_norm": 2.09609705767973,
"learning_rate": 1e-06,
"loss": 0.199,
"step": 3059
},
{
"epoch": 1.1701720841300192,
"grad_norm": 2.7151776225119324,
"learning_rate": 1e-06,
"loss": 0.0903,
"step": 3060
},
{
"epoch": 1.1705544933078393,
"grad_norm": 3.4406712555786036,
"learning_rate": 1e-06,
"loss": 0.1539,
"step": 3061
},
{
"epoch": 1.1709369024856597,
"grad_norm": 3.926922350483711,
"learning_rate": 1e-06,
"loss": 0.2428,
"step": 3062
},
{
"epoch": 1.17131931166348,
"grad_norm": 3.4846850885877734,
"learning_rate": 1e-06,
"loss": 0.0542,
"step": 3063
},
{
"epoch": 1.1717017208413,
"grad_norm": 3.7392877800199655,
"learning_rate": 1e-06,
"loss": 0.0978,
"step": 3064
},
{
"epoch": 1.1720841300191205,
"grad_norm": 5.540720961719221,
"learning_rate": 1e-06,
"loss": 0.146,
"step": 3065
},
{
"epoch": 1.1724665391969407,
"grad_norm": 6.176756500123378,
"learning_rate": 1e-06,
"loss": 0.3402,
"step": 3066
},
{
"epoch": 1.172848948374761,
"grad_norm": 5.918927508454703,
"learning_rate": 1e-06,
"loss": 0.3864,
"step": 3067
},
{
"epoch": 1.1732313575525812,
"grad_norm": 2.693373184224026,
"learning_rate": 1e-06,
"loss": 0.2767,
"step": 3068
},
{
"epoch": 1.1736137667304014,
"grad_norm": 2.91300425605188,
"learning_rate": 1e-06,
"loss": 0.3393,
"step": 3069
},
{
"epoch": 1.1739961759082218,
"grad_norm": 3.2628369270654747,
"learning_rate": 1e-06,
"loss": 0.4239,
"step": 3070
},
{
"epoch": 1.174378585086042,
"grad_norm": 4.737414703003876,
"learning_rate": 1e-06,
"loss": 0.2154,
"step": 3071
},
{
"epoch": 1.1747609942638624,
"grad_norm": 4.059744501953611,
"learning_rate": 1e-06,
"loss": 0.2917,
"step": 3072
},
{
"epoch": 1.1751434034416826,
"grad_norm": 2.408483682271091,
"learning_rate": 1e-06,
"loss": 0.2515,
"step": 3073
},
{
"epoch": 1.175525812619503,
"grad_norm": 2.8249414556057793,
"learning_rate": 1e-06,
"loss": 0.1551,
"step": 3074
},
{
"epoch": 1.1759082217973231,
"grad_norm": 3.9065458872311956,
"learning_rate": 1e-06,
"loss": 0.132,
"step": 3075
},
{
"epoch": 1.1762906309751435,
"grad_norm": 3.04849106405762,
"learning_rate": 1e-06,
"loss": 0.0795,
"step": 3076
},
{
"epoch": 1.1766730401529637,
"grad_norm": 3.563581871260127,
"learning_rate": 1e-06,
"loss": 0.0695,
"step": 3077
},
{
"epoch": 1.1770554493307839,
"grad_norm": 5.442069402788394,
"learning_rate": 1e-06,
"loss": 0.2164,
"step": 3078
},
{
"epoch": 1.1774378585086043,
"grad_norm": 4.555951998034926,
"learning_rate": 1e-06,
"loss": 0.3949,
"step": 3079
},
{
"epoch": 1.1778202676864244,
"grad_norm": 3.025138908538679,
"learning_rate": 1e-06,
"loss": 0.2971,
"step": 3080
},
{
"epoch": 1.1782026768642448,
"grad_norm": 1.8902881416798445,
"learning_rate": 1e-06,
"loss": 0.184,
"step": 3081
},
{
"epoch": 1.178585086042065,
"grad_norm": 2.9900248308968744,
"learning_rate": 1e-06,
"loss": 0.3602,
"step": 3082
},
{
"epoch": 1.1789674952198852,
"grad_norm": 2.448414550362092,
"learning_rate": 1e-06,
"loss": 0.1353,
"step": 3083
},
{
"epoch": 1.1793499043977056,
"grad_norm": 2.5123728705739317,
"learning_rate": 1e-06,
"loss": 0.2672,
"step": 3084
},
{
"epoch": 1.1797323135755258,
"grad_norm": 2.6692734930754027,
"learning_rate": 1e-06,
"loss": 0.3763,
"step": 3085
},
{
"epoch": 1.1801147227533462,
"grad_norm": 3.5069497455686953,
"learning_rate": 1e-06,
"loss": 0.2568,
"step": 3086
},
{
"epoch": 1.1804971319311663,
"grad_norm": 3.536805905025983,
"learning_rate": 1e-06,
"loss": 0.314,
"step": 3087
},
{
"epoch": 1.1808795411089865,
"grad_norm": 3.8165563048241626,
"learning_rate": 1e-06,
"loss": 0.1105,
"step": 3088
},
{
"epoch": 1.181261950286807,
"grad_norm": 2.945043414950344,
"learning_rate": 1e-06,
"loss": 0.0548,
"step": 3089
},
{
"epoch": 1.181644359464627,
"grad_norm": 4.024186681235672,
"learning_rate": 1e-06,
"loss": 0.2052,
"step": 3090
},
{
"epoch": 1.1820267686424475,
"grad_norm": 4.989335512551899,
"learning_rate": 1e-06,
"loss": 0.3828,
"step": 3091
},
{
"epoch": 1.1824091778202677,
"grad_norm": 3.8864247886636702,
"learning_rate": 1e-06,
"loss": 0.392,
"step": 3092
},
{
"epoch": 1.1827915869980878,
"grad_norm": 2.6479801469502333,
"learning_rate": 1e-06,
"loss": 0.4046,
"step": 3093
},
{
"epoch": 1.1831739961759082,
"grad_norm": 3.7650680446681766,
"learning_rate": 1e-06,
"loss": 0.3263,
"step": 3094
},
{
"epoch": 1.1835564053537284,
"grad_norm": 2.9397296965506436,
"learning_rate": 1e-06,
"loss": 0.2835,
"step": 3095
},
{
"epoch": 1.1839388145315488,
"grad_norm": 3.034743350896505,
"learning_rate": 1e-06,
"loss": 0.2192,
"step": 3096
},
{
"epoch": 1.184321223709369,
"grad_norm": 3.721951172735419,
"learning_rate": 1e-06,
"loss": 0.321,
"step": 3097
},
{
"epoch": 1.1847036328871894,
"grad_norm": 2.5948500541040307,
"learning_rate": 1e-06,
"loss": 0.3201,
"step": 3098
},
{
"epoch": 1.1850860420650096,
"grad_norm": 3.331996617914249,
"learning_rate": 1e-06,
"loss": 0.2145,
"step": 3099
},
{
"epoch": 1.1854684512428297,
"grad_norm": 3.5142993377350598,
"learning_rate": 1e-06,
"loss": 0.2495,
"step": 3100
},
{
"epoch": 1.1858508604206501,
"grad_norm": 3.116332890662486,
"learning_rate": 1e-06,
"loss": 0.1489,
"step": 3101
},
{
"epoch": 1.1862332695984703,
"grad_norm": 5.030351262350556,
"learning_rate": 1e-06,
"loss": 0.1509,
"step": 3102
},
{
"epoch": 1.1866156787762907,
"grad_norm": 5.315250178538092,
"learning_rate": 1e-06,
"loss": 0.1946,
"step": 3103
},
{
"epoch": 1.1869980879541109,
"grad_norm": 2.920293206426201,
"learning_rate": 1e-06,
"loss": 0.3817,
"step": 3104
},
{
"epoch": 1.1873804971319313,
"grad_norm": 3.344548923202473,
"learning_rate": 1e-06,
"loss": 0.4088,
"step": 3105
},
{
"epoch": 1.1877629063097515,
"grad_norm": 2.7876508116018255,
"learning_rate": 1e-06,
"loss": 0.2356,
"step": 3106
},
{
"epoch": 1.1881453154875716,
"grad_norm": 3.4527489642980744,
"learning_rate": 1e-06,
"loss": 0.3825,
"step": 3107
},
{
"epoch": 1.188527724665392,
"grad_norm": 2.055256695830084,
"learning_rate": 1e-06,
"loss": 0.3209,
"step": 3108
},
{
"epoch": 1.1889101338432122,
"grad_norm": 6.191801175298961,
"learning_rate": 1e-06,
"loss": 0.2992,
"step": 3109
},
{
"epoch": 1.1892925430210326,
"grad_norm": 2.7793039092606873,
"learning_rate": 1e-06,
"loss": 0.2882,
"step": 3110
},
{
"epoch": 1.1896749521988528,
"grad_norm": 2.7949971122931507,
"learning_rate": 1e-06,
"loss": 0.1721,
"step": 3111
},
{
"epoch": 1.190057361376673,
"grad_norm": 2.5417224273401775,
"learning_rate": 1e-06,
"loss": 0.1158,
"step": 3112
},
{
"epoch": 1.1904397705544933,
"grad_norm": 3.445079423628347,
"learning_rate": 1e-06,
"loss": 0.1185,
"step": 3113
},
{
"epoch": 1.1908221797323135,
"grad_norm": 3.27700822133397,
"learning_rate": 1e-06,
"loss": 0.0558,
"step": 3114
},
{
"epoch": 1.191204588910134,
"grad_norm": 5.881778783265607,
"learning_rate": 1e-06,
"loss": 0.0909,
"step": 3115
},
{
"epoch": 1.191586998087954,
"grad_norm": 4.455296037387607,
"learning_rate": 1e-06,
"loss": 0.1491,
"step": 3116
},
{
"epoch": 1.1919694072657743,
"grad_norm": 4.650520229976839,
"learning_rate": 1e-06,
"loss": 0.7205,
"step": 3117
},
{
"epoch": 1.1923518164435947,
"grad_norm": 2.8509326814098337,
"learning_rate": 1e-06,
"loss": 0.3414,
"step": 3118
},
{
"epoch": 1.1927342256214148,
"grad_norm": 2.75630315266608,
"learning_rate": 1e-06,
"loss": 0.2999,
"step": 3119
},
{
"epoch": 1.1931166347992352,
"grad_norm": 2.8343341593128977,
"learning_rate": 1e-06,
"loss": 0.2819,
"step": 3120
},
{
"epoch": 1.1934990439770554,
"grad_norm": 5.30408160832426,
"learning_rate": 1e-06,
"loss": 0.2805,
"step": 3121
},
{
"epoch": 1.1938814531548756,
"grad_norm": 2.828519509129359,
"learning_rate": 1e-06,
"loss": 0.1974,
"step": 3122
},
{
"epoch": 1.194263862332696,
"grad_norm": 2.3947696038206474,
"learning_rate": 1e-06,
"loss": 0.1117,
"step": 3123
},
{
"epoch": 1.1946462715105162,
"grad_norm": 3.3063217703343417,
"learning_rate": 1e-06,
"loss": 0.2637,
"step": 3124
},
{
"epoch": 1.1950286806883366,
"grad_norm": 2.5935077783936054,
"learning_rate": 1e-06,
"loss": 0.0736,
"step": 3125
},
{
"epoch": 1.1954110898661567,
"grad_norm": 3.0294580570433327,
"learning_rate": 1e-06,
"loss": 0.078,
"step": 3126
},
{
"epoch": 1.1957934990439771,
"grad_norm": 2.656613762574868,
"learning_rate": 1e-06,
"loss": 0.0651,
"step": 3127
},
{
"epoch": 1.1961759082217973,
"grad_norm": 5.002093449076198,
"learning_rate": 1e-06,
"loss": 0.2206,
"step": 3128
},
{
"epoch": 1.1965583173996177,
"grad_norm": 5.949253695830555,
"learning_rate": 1e-06,
"loss": 0.3757,
"step": 3129
},
{
"epoch": 1.1969407265774379,
"grad_norm": 3.824105780299611,
"learning_rate": 1e-06,
"loss": 0.327,
"step": 3130
},
{
"epoch": 1.197323135755258,
"grad_norm": 1.6669698757294193,
"learning_rate": 1e-06,
"loss": 0.2487,
"step": 3131
},
{
"epoch": 1.1977055449330785,
"grad_norm": 3.2204468337705974,
"learning_rate": 1e-06,
"loss": 0.2952,
"step": 3132
},
{
"epoch": 1.1980879541108986,
"grad_norm": 2.9390804827645467,
"learning_rate": 1e-06,
"loss": 0.2063,
"step": 3133
},
{
"epoch": 1.198470363288719,
"grad_norm": 3.1498462942254966,
"learning_rate": 1e-06,
"loss": 0.1814,
"step": 3134
},
{
"epoch": 1.1988527724665392,
"grad_norm": 2.3931106908873043,
"learning_rate": 1e-06,
"loss": 0.1524,
"step": 3135
},
{
"epoch": 1.1992351816443594,
"grad_norm": 2.448078577625523,
"learning_rate": 1e-06,
"loss": 0.1638,
"step": 3136
},
{
"epoch": 1.1996175908221798,
"grad_norm": 2.7783528357883305,
"learning_rate": 1e-06,
"loss": 0.107,
"step": 3137
},
{
"epoch": 1.2,
"grad_norm": 4.39841516736393,
"learning_rate": 1e-06,
"loss": 0.3795,
"step": 3138
},
{
"epoch": 1.2003824091778204,
"grad_norm": 5.160358618465096,
"learning_rate": 1e-06,
"loss": 0.122,
"step": 3139
},
{
"epoch": 1.2007648183556405,
"grad_norm": 3.327718647251247,
"learning_rate": 1e-06,
"loss": 0.0418,
"step": 3140
},
{
"epoch": 1.2011472275334607,
"grad_norm": 3.2284810640731814,
"learning_rate": 1e-06,
"loss": 0.4536,
"step": 3141
},
{
"epoch": 1.201529636711281,
"grad_norm": 4.084538470078328,
"learning_rate": 1e-06,
"loss": 0.4799,
"step": 3142
},
{
"epoch": 1.2019120458891013,
"grad_norm": 3.1492961157662314,
"learning_rate": 1e-06,
"loss": 0.3956,
"step": 3143
},
{
"epoch": 1.2022944550669217,
"grad_norm": 3.3343670831530896,
"learning_rate": 1e-06,
"loss": 0.2856,
"step": 3144
},
{
"epoch": 1.2026768642447419,
"grad_norm": 3.375294001454161,
"learning_rate": 1e-06,
"loss": 0.4036,
"step": 3145
},
{
"epoch": 1.203059273422562,
"grad_norm": 3.1989175098082954,
"learning_rate": 1e-06,
"loss": 0.1995,
"step": 3146
},
{
"epoch": 1.2034416826003824,
"grad_norm": 3.151750132885931,
"learning_rate": 1e-06,
"loss": 0.1994,
"step": 3147
},
{
"epoch": 1.2038240917782026,
"grad_norm": 2.7971436781738004,
"learning_rate": 1e-06,
"loss": 0.2323,
"step": 3148
},
{
"epoch": 1.204206500956023,
"grad_norm": 3.129228249124635,
"learning_rate": 1e-06,
"loss": 0.1711,
"step": 3149
},
{
"epoch": 1.2045889101338432,
"grad_norm": 4.379543016536247,
"learning_rate": 1e-06,
"loss": 0.2515,
"step": 3150
},
{
"epoch": 1.2049713193116636,
"grad_norm": 3.2287229827920165,
"learning_rate": 1e-06,
"loss": 0.1618,
"step": 3151
},
{
"epoch": 1.2053537284894837,
"grad_norm": 3.0254380657695057,
"learning_rate": 1e-06,
"loss": 0.1024,
"step": 3152
},
{
"epoch": 1.205736137667304,
"grad_norm": 6.945429214166259,
"learning_rate": 1e-06,
"loss": 0.3903,
"step": 3153
},
{
"epoch": 1.2061185468451243,
"grad_norm": 3.033570649920902,
"learning_rate": 1e-06,
"loss": 0.5589,
"step": 3154
},
{
"epoch": 1.2065009560229445,
"grad_norm": 4.548874877322538,
"learning_rate": 1e-06,
"loss": 0.2562,
"step": 3155
},
{
"epoch": 1.206883365200765,
"grad_norm": 2.5369299751033494,
"learning_rate": 1e-06,
"loss": 0.1815,
"step": 3156
},
{
"epoch": 1.207265774378585,
"grad_norm": 2.8103183337956694,
"learning_rate": 1e-06,
"loss": 0.3868,
"step": 3157
},
{
"epoch": 1.2076481835564055,
"grad_norm": 2.536577340740925,
"learning_rate": 1e-06,
"loss": 0.2505,
"step": 3158
},
{
"epoch": 1.2080305927342256,
"grad_norm": 2.4715991419671997,
"learning_rate": 1e-06,
"loss": 0.2468,
"step": 3159
},
{
"epoch": 1.2084130019120458,
"grad_norm": 4.598208385363217,
"learning_rate": 1e-06,
"loss": 0.2992,
"step": 3160
},
{
"epoch": 1.2087954110898662,
"grad_norm": 3.0435948639799157,
"learning_rate": 1e-06,
"loss": 0.0999,
"step": 3161
},
{
"epoch": 1.2091778202676864,
"grad_norm": 2.9515441215431655,
"learning_rate": 1e-06,
"loss": 0.1144,
"step": 3162
},
{
"epoch": 1.2095602294455068,
"grad_norm": 3.805486011720499,
"learning_rate": 1e-06,
"loss": 0.0523,
"step": 3163
},
{
"epoch": 1.209942638623327,
"grad_norm": 3.691795821043919,
"learning_rate": 1e-06,
"loss": 0.1492,
"step": 3164
},
{
"epoch": 1.2103250478011471,
"grad_norm": 3.65279127020977,
"learning_rate": 1e-06,
"loss": 0.0802,
"step": 3165
},
{
"epoch": 1.2107074569789675,
"grad_norm": 5.169597573767907,
"learning_rate": 1e-06,
"loss": 0.4806,
"step": 3166
},
{
"epoch": 1.2110898661567877,
"grad_norm": 3.912581294863535,
"learning_rate": 1e-06,
"loss": 0.5567,
"step": 3167
},
{
"epoch": 1.211472275334608,
"grad_norm": 3.827730866987371,
"learning_rate": 1e-06,
"loss": 0.4558,
"step": 3168
},
{
"epoch": 1.2118546845124283,
"grad_norm": 1.7782685609060942,
"learning_rate": 1e-06,
"loss": 0.2566,
"step": 3169
},
{
"epoch": 1.2122370936902485,
"grad_norm": 1.924411743249389,
"learning_rate": 1e-06,
"loss": 0.2081,
"step": 3170
},
{
"epoch": 1.2126195028680689,
"grad_norm": 2.906422763734687,
"learning_rate": 1e-06,
"loss": 0.2273,
"step": 3171
},
{
"epoch": 1.213001912045889,
"grad_norm": 3.692293058574239,
"learning_rate": 1e-06,
"loss": 0.2721,
"step": 3172
},
{
"epoch": 1.2133843212237094,
"grad_norm": 2.9293873544167552,
"learning_rate": 1e-06,
"loss": 0.1852,
"step": 3173
},
{
"epoch": 1.2137667304015296,
"grad_norm": 3.8968379039574175,
"learning_rate": 1e-06,
"loss": 0.2908,
"step": 3174
},
{
"epoch": 1.21414913957935,
"grad_norm": 3.956550892842321,
"learning_rate": 1e-06,
"loss": 0.1222,
"step": 3175
},
{
"epoch": 1.2145315487571702,
"grad_norm": 2.752969785508569,
"learning_rate": 1e-06,
"loss": 0.0564,
"step": 3176
},
{
"epoch": 1.2149139579349904,
"grad_norm": 3.5963030163964476,
"learning_rate": 1e-06,
"loss": 0.095,
"step": 3177
},
{
"epoch": 1.2152963671128107,
"grad_norm": 6.239211575021387,
"learning_rate": 1e-06,
"loss": 0.3205,
"step": 3178
},
{
"epoch": 1.215678776290631,
"grad_norm": 3.5378499427487347,
"learning_rate": 1e-06,
"loss": 0.3933,
"step": 3179
},
{
"epoch": 1.2160611854684513,
"grad_norm": 4.331651263344768,
"learning_rate": 1e-06,
"loss": 0.569,
"step": 3180
},
{
"epoch": 1.2164435946462715,
"grad_norm": 2.1014254500503085,
"learning_rate": 1e-06,
"loss": 0.2221,
"step": 3181
},
{
"epoch": 1.216826003824092,
"grad_norm": 2.570308511377272,
"learning_rate": 1e-06,
"loss": 0.3164,
"step": 3182
},
{
"epoch": 1.217208413001912,
"grad_norm": 2.5685503174329316,
"learning_rate": 1e-06,
"loss": 0.2736,
"step": 3183
},
{
"epoch": 1.2175908221797322,
"grad_norm": 3.2785645167553654,
"learning_rate": 1e-06,
"loss": 0.1699,
"step": 3184
},
{
"epoch": 1.2179732313575526,
"grad_norm": 2.6871196566403186,
"learning_rate": 1e-06,
"loss": 0.146,
"step": 3185
},
{
"epoch": 1.2183556405353728,
"grad_norm": 3.3899147239019376,
"learning_rate": 1e-06,
"loss": 0.1565,
"step": 3186
},
{
"epoch": 1.2187380497131932,
"grad_norm": 3.341117321595433,
"learning_rate": 1e-06,
"loss": 0.1689,
"step": 3187
},
{
"epoch": 1.2191204588910134,
"grad_norm": 4.294075126498907,
"learning_rate": 1e-06,
"loss": 0.1261,
"step": 3188
},
{
"epoch": 1.2195028680688336,
"grad_norm": 6.297185468947398,
"learning_rate": 1e-06,
"loss": 0.1083,
"step": 3189
},
{
"epoch": 1.219885277246654,
"grad_norm": 6.680371629805088,
"learning_rate": 1e-06,
"loss": 0.2221,
"step": 3190
},
{
"epoch": 1.2202676864244741,
"grad_norm": 4.298948585247168,
"learning_rate": 1e-06,
"loss": 0.4022,
"step": 3191
},
{
"epoch": 1.2206500956022945,
"grad_norm": 4.209939530543174,
"learning_rate": 1e-06,
"loss": 0.4461,
"step": 3192
},
{
"epoch": 1.2210325047801147,
"grad_norm": 4.099530514144439,
"learning_rate": 1e-06,
"loss": 0.2386,
"step": 3193
},
{
"epoch": 1.221414913957935,
"grad_norm": 2.939929201220825,
"learning_rate": 1e-06,
"loss": 0.4194,
"step": 3194
},
{
"epoch": 1.2217973231357553,
"grad_norm": 3.1781842101516435,
"learning_rate": 1e-06,
"loss": 0.3398,
"step": 3195
},
{
"epoch": 1.2221797323135755,
"grad_norm": 3.066835720820562,
"learning_rate": 1e-06,
"loss": 0.4716,
"step": 3196
},
{
"epoch": 1.2225621414913959,
"grad_norm": 3.2932464735903735,
"learning_rate": 1e-06,
"loss": 0.373,
"step": 3197
},
{
"epoch": 1.222944550669216,
"grad_norm": 3.2954179337240603,
"learning_rate": 1e-06,
"loss": 0.3699,
"step": 3198
},
{
"epoch": 1.2233269598470362,
"grad_norm": 2.894484793718795,
"learning_rate": 1e-06,
"loss": 0.1878,
"step": 3199
},
{
"epoch": 1.2237093690248566,
"grad_norm": 3.8646033081744915,
"learning_rate": 1e-06,
"loss": 0.1205,
"step": 3200
},
{
"epoch": 1.2240917782026768,
"grad_norm": 3.157835307314965,
"learning_rate": 1e-06,
"loss": 0.2107,
"step": 3201
},
{
"epoch": 1.2244741873804972,
"grad_norm": 3.8377591719695143,
"learning_rate": 1e-06,
"loss": 0.1542,
"step": 3202
},
{
"epoch": 1.2248565965583174,
"grad_norm": 8.411278791056619,
"learning_rate": 1e-06,
"loss": 0.1974,
"step": 3203
},
{
"epoch": 1.2252390057361378,
"grad_norm": 4.466116175724537,
"learning_rate": 1e-06,
"loss": 0.3859,
"step": 3204
},
{
"epoch": 1.225621414913958,
"grad_norm": 4.452325688954857,
"learning_rate": 1e-06,
"loss": 0.4961,
"step": 3205
},
{
"epoch": 1.2260038240917783,
"grad_norm": 2.7106294044633,
"learning_rate": 1e-06,
"loss": 0.3824,
"step": 3206
},
{
"epoch": 1.2263862332695985,
"grad_norm": 2.5069241955879553,
"learning_rate": 1e-06,
"loss": 0.3542,
"step": 3207
},
{
"epoch": 1.2267686424474187,
"grad_norm": 4.087994438854399,
"learning_rate": 1e-06,
"loss": 0.4534,
"step": 3208
},
{
"epoch": 1.227151051625239,
"grad_norm": 3.07792267884815,
"learning_rate": 1e-06,
"loss": 0.2163,
"step": 3209
},
{
"epoch": 1.2275334608030593,
"grad_norm": 2.741479505123859,
"learning_rate": 1e-06,
"loss": 0.1893,
"step": 3210
},
{
"epoch": 1.2279158699808796,
"grad_norm": 2.9077849897736474,
"learning_rate": 1e-06,
"loss": 0.2899,
"step": 3211
},
{
"epoch": 1.2282982791586998,
"grad_norm": 3.499013352971792,
"learning_rate": 1e-06,
"loss": 0.1566,
"step": 3212
},
{
"epoch": 1.22868068833652,
"grad_norm": 2.333498086108299,
"learning_rate": 1e-06,
"loss": 0.0706,
"step": 3213
},
{
"epoch": 1.2290630975143404,
"grad_norm": 5.4533163217467315,
"learning_rate": 1e-06,
"loss": 0.0639,
"step": 3214
},
{
"epoch": 1.2294455066921606,
"grad_norm": 5.173192343380095,
"learning_rate": 1e-06,
"loss": 0.1286,
"step": 3215
},
{
"epoch": 1.229827915869981,
"grad_norm": 7.6937607493197655,
"learning_rate": 1e-06,
"loss": 0.3088,
"step": 3216
},
{
"epoch": 1.2302103250478011,
"grad_norm": 2.636512488315521,
"learning_rate": 1e-06,
"loss": 0.3282,
"step": 3217
},
{
"epoch": 1.2305927342256213,
"grad_norm": 3.1988446176420298,
"learning_rate": 1e-06,
"loss": 0.5207,
"step": 3218
},
{
"epoch": 1.2309751434034417,
"grad_norm": 2.15027620182632,
"learning_rate": 1e-06,
"loss": 0.2561,
"step": 3219
},
{
"epoch": 1.231357552581262,
"grad_norm": 2.987927943268552,
"learning_rate": 1e-06,
"loss": 0.2925,
"step": 3220
},
{
"epoch": 1.2317399617590823,
"grad_norm": 4.174506910981116,
"learning_rate": 1e-06,
"loss": 0.4875,
"step": 3221
},
{
"epoch": 1.2321223709369025,
"grad_norm": 2.3061472262508063,
"learning_rate": 1e-06,
"loss": 0.3425,
"step": 3222
},
{
"epoch": 1.2325047801147226,
"grad_norm": 2.5689735519514834,
"learning_rate": 1e-06,
"loss": 0.1534,
"step": 3223
},
{
"epoch": 1.232887189292543,
"grad_norm": 3.447316804241193,
"learning_rate": 1e-06,
"loss": 0.1273,
"step": 3224
},
{
"epoch": 1.2332695984703632,
"grad_norm": 3.474353285514223,
"learning_rate": 1e-06,
"loss": 0.1209,
"step": 3225
},
{
"epoch": 1.2336520076481836,
"grad_norm": 2.7847654434955316,
"learning_rate": 1e-06,
"loss": 0.0734,
"step": 3226
},
{
"epoch": 1.2340344168260038,
"grad_norm": 7.495555069151165,
"learning_rate": 1e-06,
"loss": 0.1253,
"step": 3227
},
{
"epoch": 1.2344168260038242,
"grad_norm": 6.239125366047068,
"learning_rate": 1e-06,
"loss": 0.2443,
"step": 3228
},
{
"epoch": 1.2347992351816444,
"grad_norm": 4.006409516632794,
"learning_rate": 1e-06,
"loss": 0.2613,
"step": 3229
},
{
"epoch": 1.2351816443594645,
"grad_norm": 3.460577056921942,
"learning_rate": 1e-06,
"loss": 0.3549,
"step": 3230
},
{
"epoch": 1.235564053537285,
"grad_norm": 2.0768008495023427,
"learning_rate": 1e-06,
"loss": 0.2616,
"step": 3231
},
{
"epoch": 1.235946462715105,
"grad_norm": 3.5882055884651836,
"learning_rate": 1e-06,
"loss": 0.2864,
"step": 3232
},
{
"epoch": 1.2363288718929255,
"grad_norm": 3.511522943216704,
"learning_rate": 1e-06,
"loss": 0.3731,
"step": 3233
},
{
"epoch": 1.2367112810707457,
"grad_norm": 15.408540690821539,
"learning_rate": 1e-06,
"loss": 0.2628,
"step": 3234
},
{
"epoch": 1.237093690248566,
"grad_norm": 3.8764653050309117,
"learning_rate": 1e-06,
"loss": 0.1392,
"step": 3235
},
{
"epoch": 1.2374760994263863,
"grad_norm": 2.5612972507242073,
"learning_rate": 1e-06,
"loss": 0.1312,
"step": 3236
},
{
"epoch": 1.2378585086042064,
"grad_norm": 2.7494724808096356,
"learning_rate": 1e-06,
"loss": 0.2313,
"step": 3237
},
{
"epoch": 1.2382409177820268,
"grad_norm": 3.3998463483760437,
"learning_rate": 1e-06,
"loss": 0.2531,
"step": 3238
},
{
"epoch": 1.238623326959847,
"grad_norm": 3.4332080750316316,
"learning_rate": 1e-06,
"loss": 0.0872,
"step": 3239
},
{
"epoch": 1.2390057361376674,
"grad_norm": 3.708626439087289,
"learning_rate": 1e-06,
"loss": 0.0635,
"step": 3240
},
{
"epoch": 1.2393881453154876,
"grad_norm": 7.655797579107328,
"learning_rate": 1e-06,
"loss": 0.3263,
"step": 3241
},
{
"epoch": 1.2397705544933078,
"grad_norm": 3.57863483378223,
"learning_rate": 1e-06,
"loss": 0.3405,
"step": 3242
},
{
"epoch": 1.2401529636711282,
"grad_norm": 3.602513781381543,
"learning_rate": 1e-06,
"loss": 0.4766,
"step": 3243
},
{
"epoch": 1.2405353728489483,
"grad_norm": 2.966494375933703,
"learning_rate": 1e-06,
"loss": 0.3667,
"step": 3244
},
{
"epoch": 1.2409177820267687,
"grad_norm": 3.3422614376098685,
"learning_rate": 1e-06,
"loss": 0.3988,
"step": 3245
},
{
"epoch": 1.241300191204589,
"grad_norm": 2.5268776889874465,
"learning_rate": 1e-06,
"loss": 0.1166,
"step": 3246
},
{
"epoch": 1.241682600382409,
"grad_norm": 4.3453519872183275,
"learning_rate": 1e-06,
"loss": 0.466,
"step": 3247
},
{
"epoch": 1.2420650095602295,
"grad_norm": 3.199253179902874,
"learning_rate": 1e-06,
"loss": 0.1567,
"step": 3248
},
{
"epoch": 1.2424474187380496,
"grad_norm": 2.5782818370855645,
"learning_rate": 1e-06,
"loss": 0.2876,
"step": 3249
},
{
"epoch": 1.24282982791587,
"grad_norm": 2.9079991481102874,
"learning_rate": 1e-06,
"loss": 0.2807,
"step": 3250
},
{
"epoch": 1.2432122370936902,
"grad_norm": 3.247128905721344,
"learning_rate": 1e-06,
"loss": 0.1605,
"step": 3251
},
{
"epoch": 1.2435946462715104,
"grad_norm": 3.7079759204086757,
"learning_rate": 1e-06,
"loss": 0.0931,
"step": 3252
},
{
"epoch": 1.2439770554493308,
"grad_norm": 5.32563826626586,
"learning_rate": 1e-06,
"loss": 0.1665,
"step": 3253
},
{
"epoch": 1.244359464627151,
"grad_norm": 4.559249347003769,
"learning_rate": 1e-06,
"loss": 0.4295,
"step": 3254
},
{
"epoch": 1.2447418738049714,
"grad_norm": 2.8274031081323483,
"learning_rate": 1e-06,
"loss": 0.3379,
"step": 3255
},
{
"epoch": 1.2451242829827915,
"grad_norm": 2.539665925140947,
"learning_rate": 1e-06,
"loss": 0.2315,
"step": 3256
},
{
"epoch": 1.245506692160612,
"grad_norm": 2.1872441823421767,
"learning_rate": 1e-06,
"loss": 0.2804,
"step": 3257
},
{
"epoch": 1.2458891013384321,
"grad_norm": 2.984422293882462,
"learning_rate": 1e-06,
"loss": 0.2912,
"step": 3258
},
{
"epoch": 1.2462715105162525,
"grad_norm": 3.5871001708627017,
"learning_rate": 1e-06,
"loss": 0.2893,
"step": 3259
},
{
"epoch": 1.2466539196940727,
"grad_norm": 2.5731002392062807,
"learning_rate": 1e-06,
"loss": 0.1797,
"step": 3260
},
{
"epoch": 1.2470363288718929,
"grad_norm": 2.6045865648618096,
"learning_rate": 1e-06,
"loss": 0.1515,
"step": 3261
},
{
"epoch": 1.2474187380497133,
"grad_norm": 2.938303898321132,
"learning_rate": 1e-06,
"loss": 0.1139,
"step": 3262
},
{
"epoch": 1.2478011472275334,
"grad_norm": 3.1936588998481197,
"learning_rate": 1e-06,
"loss": 0.0828,
"step": 3263
},
{
"epoch": 1.2481835564053538,
"grad_norm": 3.4553206009902238,
"learning_rate": 1e-06,
"loss": 0.1139,
"step": 3264
},
{
"epoch": 1.248565965583174,
"grad_norm": 3.466891779070712,
"learning_rate": 1e-06,
"loss": 0.0688,
"step": 3265
},
{
"epoch": 1.2489483747609942,
"grad_norm": 5.993548421483226,
"learning_rate": 1e-06,
"loss": 0.4569,
"step": 3266
},
{
"epoch": 1.2493307839388146,
"grad_norm": 3.3922338931488256,
"learning_rate": 1e-06,
"loss": 0.6279,
"step": 3267
},
{
"epoch": 1.2497131931166348,
"grad_norm": 2.6383454906627724,
"learning_rate": 1e-06,
"loss": 0.2484,
"step": 3268
},
{
"epoch": 1.2500956022944552,
"grad_norm": 4.660954908184366,
"learning_rate": 1e-06,
"loss": 0.3153,
"step": 3269
},
{
"epoch": 1.2504780114722753,
"grad_norm": 3.2982163435081064,
"learning_rate": 1e-06,
"loss": 0.2657,
"step": 3270
},
{
"epoch": 1.2508604206500955,
"grad_norm": 3.52731657809821,
"learning_rate": 1e-06,
"loss": 0.3095,
"step": 3271
},
{
"epoch": 1.251242829827916,
"grad_norm": 2.9784126839280836,
"learning_rate": 1e-06,
"loss": 0.1788,
"step": 3272
},
{
"epoch": 1.251625239005736,
"grad_norm": 3.3139081157315933,
"learning_rate": 1e-06,
"loss": 0.1776,
"step": 3273
},
{
"epoch": 1.2520076481835565,
"grad_norm": 3.064465106383923,
"learning_rate": 1e-06,
"loss": 0.1641,
"step": 3274
},
{
"epoch": 1.2523900573613767,
"grad_norm": 3.0953936593643636,
"learning_rate": 1e-06,
"loss": 0.0889,
"step": 3275
},
{
"epoch": 1.2527724665391968,
"grad_norm": 2.613790738970916,
"learning_rate": 1e-06,
"loss": 0.0905,
"step": 3276
},
{
"epoch": 1.2531548757170172,
"grad_norm": 2.995440037518959,
"learning_rate": 1e-06,
"loss": 0.063,
"step": 3277
},
{
"epoch": 1.2535372848948374,
"grad_norm": 4.153778029423889,
"learning_rate": 1e-06,
"loss": 0.1795,
"step": 3278
},
{
"epoch": 1.2539196940726578,
"grad_norm": 3.231124497711958,
"learning_rate": 1e-06,
"loss": 0.3398,
"step": 3279
},
{
"epoch": 1.254302103250478,
"grad_norm": 2.467408793733926,
"learning_rate": 1e-06,
"loss": 0.3896,
"step": 3280
},
{
"epoch": 1.2546845124282981,
"grad_norm": 2.5752837274770437,
"learning_rate": 1e-06,
"loss": 0.3662,
"step": 3281
},
{
"epoch": 1.2550669216061185,
"grad_norm": 3.266375587755585,
"learning_rate": 1e-06,
"loss": 0.3,
"step": 3282
},
{
"epoch": 1.255449330783939,
"grad_norm": 3.6222765985423666,
"learning_rate": 1e-06,
"loss": 0.2978,
"step": 3283
},
{
"epoch": 1.2558317399617591,
"grad_norm": 4.070443705898735,
"learning_rate": 1e-06,
"loss": 0.2446,
"step": 3284
},
{
"epoch": 1.2562141491395793,
"grad_norm": 2.598440960980999,
"learning_rate": 1e-06,
"loss": 0.1429,
"step": 3285
},
{
"epoch": 1.2565965583173997,
"grad_norm": 2.6561142830678044,
"learning_rate": 1e-06,
"loss": 0.1296,
"step": 3286
},
{
"epoch": 1.2569789674952199,
"grad_norm": 2.938971293046691,
"learning_rate": 1e-06,
"loss": 0.1215,
"step": 3287
},
{
"epoch": 1.2573613766730403,
"grad_norm": 4.553445270404203,
"learning_rate": 1e-06,
"loss": 0.2336,
"step": 3288
},
{
"epoch": 1.2577437858508604,
"grad_norm": 3.817733674357759,
"learning_rate": 1e-06,
"loss": 0.1953,
"step": 3289
},
{
"epoch": 1.2581261950286806,
"grad_norm": 4.205942110353462,
"learning_rate": 1e-06,
"loss": 0.1186,
"step": 3290
},
{
"epoch": 1.258508604206501,
"grad_norm": 5.425228961384696,
"learning_rate": 1e-06,
"loss": 0.29,
"step": 3291
},
{
"epoch": 1.2588910133843212,
"grad_norm": 3.728292707330749,
"learning_rate": 1e-06,
"loss": 0.3458,
"step": 3292
},
{
"epoch": 1.2592734225621416,
"grad_norm": 4.7855711714243245,
"learning_rate": 1e-06,
"loss": 0.7093,
"step": 3293
},
{
"epoch": 1.2596558317399618,
"grad_norm": 3.4524078314851363,
"learning_rate": 1e-06,
"loss": 0.3338,
"step": 3294
},
{
"epoch": 1.260038240917782,
"grad_norm": 4.576692011893362,
"learning_rate": 1e-06,
"loss": 0.5975,
"step": 3295
},
{
"epoch": 1.2604206500956023,
"grad_norm": 3.2486478486977677,
"learning_rate": 1e-06,
"loss": 0.2046,
"step": 3296
},
{
"epoch": 1.2608030592734225,
"grad_norm": 2.8974439866689035,
"learning_rate": 1e-06,
"loss": 0.1111,
"step": 3297
},
{
"epoch": 1.261185468451243,
"grad_norm": 2.3431041590632984,
"learning_rate": 1e-06,
"loss": 0.0771,
"step": 3298
},
{
"epoch": 1.261567877629063,
"grad_norm": 3.0383392493238692,
"learning_rate": 1e-06,
"loss": 0.2024,
"step": 3299
},
{
"epoch": 1.2619502868068833,
"grad_norm": 2.949298296733519,
"learning_rate": 1e-06,
"loss": 0.1405,
"step": 3300
},
{
"epoch": 1.2623326959847037,
"grad_norm": 3.538927088287955,
"learning_rate": 1e-06,
"loss": 0.1302,
"step": 3301
},
{
"epoch": 1.2627151051625238,
"grad_norm": 3.601970260607489,
"learning_rate": 1e-06,
"loss": 0.0826,
"step": 3302
},
{
"epoch": 1.2630975143403442,
"grad_norm": 5.1360362248554425,
"learning_rate": 1e-06,
"loss": 0.1543,
"step": 3303
},
{
"epoch": 1.2634799235181644,
"grad_norm": 4.114321672191617,
"learning_rate": 1e-06,
"loss": 0.2817,
"step": 3304
},
{
"epoch": 1.2638623326959846,
"grad_norm": 2.8486934076088635,
"learning_rate": 1e-06,
"loss": 0.2746,
"step": 3305
},
{
"epoch": 1.264244741873805,
"grad_norm": 2.444121927120025,
"learning_rate": 1e-06,
"loss": 0.2642,
"step": 3306
},
{
"epoch": 1.2646271510516254,
"grad_norm": 3.2443438976709884,
"learning_rate": 1e-06,
"loss": 0.6061,
"step": 3307
},
{
"epoch": 1.2650095602294456,
"grad_norm": 2.604470492123503,
"learning_rate": 1e-06,
"loss": 0.3553,
"step": 3308
},
{
"epoch": 1.2653919694072657,
"grad_norm": 3.1613936106023286,
"learning_rate": 1e-06,
"loss": 0.4285,
"step": 3309
},
{
"epoch": 1.2657743785850861,
"grad_norm": 2.5794572566869,
"learning_rate": 1e-06,
"loss": 0.1728,
"step": 3310
},
{
"epoch": 1.2661567877629063,
"grad_norm": 2.5218372290145044,
"learning_rate": 1e-06,
"loss": 0.1966,
"step": 3311
},
{
"epoch": 1.2665391969407267,
"grad_norm": 4.796943390697319,
"learning_rate": 1e-06,
"loss": 0.1027,
"step": 3312
},
{
"epoch": 1.2669216061185469,
"grad_norm": 2.730988503514679,
"learning_rate": 1e-06,
"loss": 0.1398,
"step": 3313
},
{
"epoch": 1.267304015296367,
"grad_norm": 3.111253487265852,
"learning_rate": 1e-06,
"loss": 0.0879,
"step": 3314
},
{
"epoch": 1.2676864244741874,
"grad_norm": 6.7800027429805825,
"learning_rate": 1e-06,
"loss": 0.156,
"step": 3315
},
{
"epoch": 1.2680688336520076,
"grad_norm": 7.963474337432703,
"learning_rate": 1e-06,
"loss": 0.495,
"step": 3316
},
{
"epoch": 1.268451242829828,
"grad_norm": 3.024389151697539,
"learning_rate": 1e-06,
"loss": 0.3465,
"step": 3317
},
{
"epoch": 1.2688336520076482,
"grad_norm": 2.860452084323208,
"learning_rate": 1e-06,
"loss": 0.3293,
"step": 3318
},
{
"epoch": 1.2692160611854684,
"grad_norm": 3.6980028742381386,
"learning_rate": 1e-06,
"loss": 0.2159,
"step": 3319
},
{
"epoch": 1.2695984703632888,
"grad_norm": 2.303032335671375,
"learning_rate": 1e-06,
"loss": 0.4376,
"step": 3320
},
{
"epoch": 1.269980879541109,
"grad_norm": 4.182387234113338,
"learning_rate": 1e-06,
"loss": 0.2782,
"step": 3321
},
{
"epoch": 1.2703632887189293,
"grad_norm": 2.692626676421947,
"learning_rate": 1e-06,
"loss": 0.1394,
"step": 3322
},
{
"epoch": 1.2707456978967495,
"grad_norm": 3.912974924141223,
"learning_rate": 1e-06,
"loss": 0.1964,
"step": 3323
},
{
"epoch": 1.2711281070745697,
"grad_norm": 3.876724474720001,
"learning_rate": 1e-06,
"loss": 0.1456,
"step": 3324
},
{
"epoch": 1.27151051625239,
"grad_norm": 4.701569055848703,
"learning_rate": 1e-06,
"loss": 0.2936,
"step": 3325
},
{
"epoch": 1.2718929254302103,
"grad_norm": 2.748696972178061,
"learning_rate": 1e-06,
"loss": 0.1832,
"step": 3326
},
{
"epoch": 1.2722753346080307,
"grad_norm": 3.8873667114565897,
"learning_rate": 1e-06,
"loss": 0.0835,
"step": 3327
},
{
"epoch": 1.2726577437858508,
"grad_norm": 5.0648131560708345,
"learning_rate": 1e-06,
"loss": 0.1069,
"step": 3328
},
{
"epoch": 1.273040152963671,
"grad_norm": 3.3261159644632983,
"learning_rate": 1e-06,
"loss": 0.3947,
"step": 3329
},
{
"epoch": 1.2734225621414914,
"grad_norm": 3.5728602428077467,
"learning_rate": 1e-06,
"loss": 0.5163,
"step": 3330
},
{
"epoch": 1.2738049713193116,
"grad_norm": 2.1417360415769147,
"learning_rate": 1e-06,
"loss": 0.2434,
"step": 3331
},
{
"epoch": 1.274187380497132,
"grad_norm": 2.986434465589483,
"learning_rate": 1e-06,
"loss": 0.2235,
"step": 3332
},
{
"epoch": 1.2745697896749522,
"grad_norm": 3.243911762650881,
"learning_rate": 1e-06,
"loss": 0.224,
"step": 3333
},
{
"epoch": 1.2749521988527723,
"grad_norm": 3.4212353252170034,
"learning_rate": 1e-06,
"loss": 0.2809,
"step": 3334
},
{
"epoch": 1.2753346080305927,
"grad_norm": 3.4021436104668235,
"learning_rate": 1e-06,
"loss": 0.181,
"step": 3335
},
{
"epoch": 1.2757170172084131,
"grad_norm": 3.3884305400824406,
"learning_rate": 1e-06,
"loss": 0.1349,
"step": 3336
},
{
"epoch": 1.2760994263862333,
"grad_norm": 2.9425503758710017,
"learning_rate": 1e-06,
"loss": 0.129,
"step": 3337
},
{
"epoch": 1.2764818355640535,
"grad_norm": 4.610561205285702,
"learning_rate": 1e-06,
"loss": 0.3488,
"step": 3338
},
{
"epoch": 1.2768642447418739,
"grad_norm": 4.070022426234139,
"learning_rate": 1e-06,
"loss": 0.0708,
"step": 3339
},
{
"epoch": 1.277246653919694,
"grad_norm": 3.9333845156771727,
"learning_rate": 1e-06,
"loss": 0.0555,
"step": 3340
},
{
"epoch": 1.2776290630975145,
"grad_norm": 5.3807590592963015,
"learning_rate": 1e-06,
"loss": 0.2507,
"step": 3341
},
{
"epoch": 1.2780114722753346,
"grad_norm": 3.451388928118952,
"learning_rate": 1e-06,
"loss": 0.3081,
"step": 3342
},
{
"epoch": 1.2783938814531548,
"grad_norm": 2.7262223258312663,
"learning_rate": 1e-06,
"loss": 0.2197,
"step": 3343
},
{
"epoch": 1.2787762906309752,
"grad_norm": 2.927656848331449,
"learning_rate": 1e-06,
"loss": 0.259,
"step": 3344
},
{
"epoch": 1.2791586998087954,
"grad_norm": 2.279239134069402,
"learning_rate": 1e-06,
"loss": 0.3078,
"step": 3345
},
{
"epoch": 1.2795411089866158,
"grad_norm": 3.1295168467792123,
"learning_rate": 1e-06,
"loss": 0.3413,
"step": 3346
},
{
"epoch": 1.279923518164436,
"grad_norm": 4.040598357806839,
"learning_rate": 1e-06,
"loss": 0.3243,
"step": 3347
},
{
"epoch": 1.2803059273422561,
"grad_norm": 3.267846333041392,
"learning_rate": 1e-06,
"loss": 0.1141,
"step": 3348
},
{
"epoch": 1.2806883365200765,
"grad_norm": 2.120686135370677,
"learning_rate": 1e-06,
"loss": 0.1233,
"step": 3349
},
{
"epoch": 1.2810707456978967,
"grad_norm": 4.662781505932946,
"learning_rate": 1e-06,
"loss": 0.2845,
"step": 3350
},
{
"epoch": 1.281453154875717,
"grad_norm": 2.194854797218163,
"learning_rate": 1e-06,
"loss": 0.0676,
"step": 3351
},
{
"epoch": 1.2818355640535373,
"grad_norm": 3.776868567416281,
"learning_rate": 1e-06,
"loss": 0.1338,
"step": 3352
},
{
"epoch": 1.2822179732313574,
"grad_norm": 5.714445568301108,
"learning_rate": 1e-06,
"loss": 0.2475,
"step": 3353
},
{
"epoch": 1.2826003824091778,
"grad_norm": 6.035578148177992,
"learning_rate": 1e-06,
"loss": 0.509,
"step": 3354
},
{
"epoch": 1.282982791586998,
"grad_norm": 4.109563902947867,
"learning_rate": 1e-06,
"loss": 0.4684,
"step": 3355
},
{
"epoch": 1.2833652007648184,
"grad_norm": 3.3846014987887463,
"learning_rate": 1e-06,
"loss": 0.1594,
"step": 3356
},
{
"epoch": 1.2837476099426386,
"grad_norm": 3.190726349030799,
"learning_rate": 1e-06,
"loss": 0.3597,
"step": 3357
},
{
"epoch": 1.2841300191204588,
"grad_norm": 3.3777993919562195,
"learning_rate": 1e-06,
"loss": 0.2764,
"step": 3358
},
{
"epoch": 1.2845124282982792,
"grad_norm": 4.926388563188322,
"learning_rate": 1e-06,
"loss": 0.2072,
"step": 3359
},
{
"epoch": 1.2848948374760996,
"grad_norm": 2.6804195549635703,
"learning_rate": 1e-06,
"loss": 0.1778,
"step": 3360
},
{
"epoch": 1.2852772466539197,
"grad_norm": 2.942750499502011,
"learning_rate": 1e-06,
"loss": 0.1543,
"step": 3361
},
{
"epoch": 1.28565965583174,
"grad_norm": 3.2003028905094935,
"learning_rate": 1e-06,
"loss": 0.2225,
"step": 3362
},
{
"epoch": 1.2860420650095603,
"grad_norm": 2.9642325312822067,
"learning_rate": 1e-06,
"loss": 0.1231,
"step": 3363
},
{
"epoch": 1.2864244741873805,
"grad_norm": 4.029409062150022,
"learning_rate": 1e-06,
"loss": 0.0952,
"step": 3364
},
{
"epoch": 1.2868068833652009,
"grad_norm": 3.274000315205367,
"learning_rate": 1e-06,
"loss": 0.0935,
"step": 3365
},
{
"epoch": 1.287189292543021,
"grad_norm": 4.361271490129927,
"learning_rate": 1e-06,
"loss": 0.374,
"step": 3366
},
{
"epoch": 1.2875717017208412,
"grad_norm": 3.3631651323324023,
"learning_rate": 1e-06,
"loss": 0.1535,
"step": 3367
},
{
"epoch": 1.2879541108986616,
"grad_norm": 4.472950275201895,
"learning_rate": 1e-06,
"loss": 0.6948,
"step": 3368
},
{
"epoch": 1.2883365200764818,
"grad_norm": 3.383041409330677,
"learning_rate": 1e-06,
"loss": 0.2226,
"step": 3369
},
{
"epoch": 1.2887189292543022,
"grad_norm": 3.066936782329665,
"learning_rate": 1e-06,
"loss": 0.2695,
"step": 3370
},
{
"epoch": 1.2891013384321224,
"grad_norm": 3.4226682257426444,
"learning_rate": 1e-06,
"loss": 0.2786,
"step": 3371
},
{
"epoch": 1.2894837476099426,
"grad_norm": 3.862396169705139,
"learning_rate": 1e-06,
"loss": 0.3995,
"step": 3372
},
{
"epoch": 1.289866156787763,
"grad_norm": 2.3260833185284504,
"learning_rate": 1e-06,
"loss": 0.1609,
"step": 3373
},
{
"epoch": 1.2902485659655831,
"grad_norm": 3.2342442287969453,
"learning_rate": 1e-06,
"loss": 0.1511,
"step": 3374
},
{
"epoch": 1.2906309751434035,
"grad_norm": 4.072343842923876,
"learning_rate": 1e-06,
"loss": 0.3727,
"step": 3375
},
{
"epoch": 1.2910133843212237,
"grad_norm": 3.900304435194107,
"learning_rate": 1e-06,
"loss": 0.1991,
"step": 3376
},
{
"epoch": 1.2913957934990439,
"grad_norm": 4.229406841705585,
"learning_rate": 1e-06,
"loss": 0.1204,
"step": 3377
},
{
"epoch": 1.2917782026768643,
"grad_norm": 5.929072774151254,
"learning_rate": 1e-06,
"loss": 0.2068,
"step": 3378
},
{
"epoch": 1.2921606118546844,
"grad_norm": 5.857656974691698,
"learning_rate": 1e-06,
"loss": 0.3528,
"step": 3379
},
{
"epoch": 1.2925430210325048,
"grad_norm": 3.4854770166462634,
"learning_rate": 1e-06,
"loss": 0.3787,
"step": 3380
},
{
"epoch": 1.292925430210325,
"grad_norm": 2.4692164113285253,
"learning_rate": 1e-06,
"loss": 0.2829,
"step": 3381
},
{
"epoch": 1.2933078393881452,
"grad_norm": 2.5575743063892857,
"learning_rate": 1e-06,
"loss": 0.2287,
"step": 3382
},
{
"epoch": 1.2936902485659656,
"grad_norm": 2.9849181954118578,
"learning_rate": 1e-06,
"loss": 0.2854,
"step": 3383
},
{
"epoch": 1.2940726577437858,
"grad_norm": 3.009305826027895,
"learning_rate": 1e-06,
"loss": 0.1549,
"step": 3384
},
{
"epoch": 1.2944550669216062,
"grad_norm": 2.844648785089339,
"learning_rate": 1e-06,
"loss": 0.2272,
"step": 3385
},
{
"epoch": 1.2948374760994263,
"grad_norm": 2.9873070818170366,
"learning_rate": 1e-06,
"loss": 0.1596,
"step": 3386
},
{
"epoch": 1.2952198852772465,
"grad_norm": 3.7417993523421957,
"learning_rate": 1e-06,
"loss": 0.2825,
"step": 3387
},
{
"epoch": 1.295602294455067,
"grad_norm": 3.6106958655900474,
"learning_rate": 1e-06,
"loss": 0.1479,
"step": 3388
},
{
"epoch": 1.2959847036328873,
"grad_norm": 5.045172720184043,
"learning_rate": 1e-06,
"loss": 0.1158,
"step": 3389
},
{
"epoch": 1.2963671128107075,
"grad_norm": 3.4616433967269815,
"learning_rate": 1e-06,
"loss": 0.0986,
"step": 3390
},
{
"epoch": 1.2967495219885277,
"grad_norm": 6.064327004022685,
"learning_rate": 1e-06,
"loss": 0.4425,
"step": 3391
},
{
"epoch": 1.297131931166348,
"grad_norm": 4.174517876661145,
"learning_rate": 1e-06,
"loss": 0.3251,
"step": 3392
},
{
"epoch": 1.2975143403441682,
"grad_norm": 3.382832234173097,
"learning_rate": 1e-06,
"loss": 0.2373,
"step": 3393
},
{
"epoch": 1.2978967495219886,
"grad_norm": 2.4883105696982386,
"learning_rate": 1e-06,
"loss": 0.1699,
"step": 3394
},
{
"epoch": 1.2982791586998088,
"grad_norm": 3.10111692672774,
"learning_rate": 1e-06,
"loss": 0.2498,
"step": 3395
},
{
"epoch": 1.298661567877629,
"grad_norm": 3.279655577789128,
"learning_rate": 1e-06,
"loss": 0.3954,
"step": 3396
},
{
"epoch": 1.2990439770554494,
"grad_norm": 3.242066898171103,
"learning_rate": 1e-06,
"loss": 0.3368,
"step": 3397
},
{
"epoch": 1.2994263862332696,
"grad_norm": 3.476942954516217,
"learning_rate": 1e-06,
"loss": 0.1567,
"step": 3398
},
{
"epoch": 1.29980879541109,
"grad_norm": 3.3651740123540197,
"learning_rate": 1e-06,
"loss": 0.1099,
"step": 3399
},
{
"epoch": 1.3001912045889101,
"grad_norm": 5.068972182666567,
"learning_rate": 1e-06,
"loss": 0.2837,
"step": 3400
},
{
"epoch": 1.3005736137667303,
"grad_norm": 2.556185874974078,
"learning_rate": 1e-06,
"loss": 0.0851,
"step": 3401
},
{
"epoch": 1.3009560229445507,
"grad_norm": 4.363005845437778,
"learning_rate": 1e-06,
"loss": 0.0745,
"step": 3402
},
{
"epoch": 1.3013384321223709,
"grad_norm": 6.656372355737953,
"learning_rate": 1e-06,
"loss": 0.1691,
"step": 3403
},
{
"epoch": 1.3017208413001913,
"grad_norm": 4.121266380418662,
"learning_rate": 1e-06,
"loss": 0.4311,
"step": 3404
},
{
"epoch": 1.3021032504780115,
"grad_norm": 2.798558075062855,
"learning_rate": 1e-06,
"loss": 0.2609,
"step": 3405
},
{
"epoch": 1.3024856596558316,
"grad_norm": 1.912774411629172,
"learning_rate": 1e-06,
"loss": 0.1919,
"step": 3406
},
{
"epoch": 1.302868068833652,
"grad_norm": 2.9925983037394017,
"learning_rate": 1e-06,
"loss": 0.2651,
"step": 3407
},
{
"epoch": 1.3032504780114722,
"grad_norm": 3.340020069130547,
"learning_rate": 1e-06,
"loss": 0.3107,
"step": 3408
},
{
"epoch": 1.3036328871892926,
"grad_norm": 3.84055772665031,
"learning_rate": 1e-06,
"loss": 0.2469,
"step": 3409
},
{
"epoch": 1.3040152963671128,
"grad_norm": 3.473565959742894,
"learning_rate": 1e-06,
"loss": 0.2675,
"step": 3410
},
{
"epoch": 1.304397705544933,
"grad_norm": 2.653459137453321,
"learning_rate": 1e-06,
"loss": 0.1012,
"step": 3411
},
{
"epoch": 1.3047801147227533,
"grad_norm": 2.996474101407398,
"learning_rate": 1e-06,
"loss": 0.1056,
"step": 3412
},
{
"epoch": 1.3051625239005737,
"grad_norm": 2.8112490097050937,
"learning_rate": 1e-06,
"loss": 0.0976,
"step": 3413
},
{
"epoch": 1.305544933078394,
"grad_norm": 3.459583439530436,
"learning_rate": 1e-06,
"loss": 0.0902,
"step": 3414
},
{
"epoch": 1.305927342256214,
"grad_norm": 6.510320539654921,
"learning_rate": 1e-06,
"loss": 0.2568,
"step": 3415
},
{
"epoch": 1.3063097514340345,
"grad_norm": 5.102605318798301,
"learning_rate": 1e-06,
"loss": 0.4318,
"step": 3416
},
{
"epoch": 1.3066921606118547,
"grad_norm": 3.021725191902132,
"learning_rate": 1e-06,
"loss": 0.2025,
"step": 3417
},
{
"epoch": 1.307074569789675,
"grad_norm": 3.7599091896291608,
"learning_rate": 1e-06,
"loss": 0.2958,
"step": 3418
},
{
"epoch": 1.3074569789674952,
"grad_norm": 3.176630523598782,
"learning_rate": 1e-06,
"loss": 0.2994,
"step": 3419
},
{
"epoch": 1.3078393881453154,
"grad_norm": 2.505285683519663,
"learning_rate": 1e-06,
"loss": 0.205,
"step": 3420
},
{
"epoch": 1.3082217973231358,
"grad_norm": 2.6696898688174446,
"learning_rate": 1e-06,
"loss": 0.2454,
"step": 3421
},
{
"epoch": 1.308604206500956,
"grad_norm": 1.8595218760718364,
"learning_rate": 1e-06,
"loss": 0.1584,
"step": 3422
},
{
"epoch": 1.3089866156787764,
"grad_norm": 4.18426183500529,
"learning_rate": 1e-06,
"loss": 0.2838,
"step": 3423
},
{
"epoch": 1.3093690248565966,
"grad_norm": 2.7474685635213083,
"learning_rate": 1e-06,
"loss": 0.1561,
"step": 3424
},
{
"epoch": 1.3097514340344167,
"grad_norm": 3.43407949715755,
"learning_rate": 1e-06,
"loss": 0.1478,
"step": 3425
},
{
"epoch": 1.3101338432122371,
"grad_norm": 2.822721429996645,
"learning_rate": 1e-06,
"loss": 0.0598,
"step": 3426
},
{
"epoch": 1.3105162523900573,
"grad_norm": 4.008866258962873,
"learning_rate": 1e-06,
"loss": 0.1668,
"step": 3427
},
{
"epoch": 1.3108986615678777,
"grad_norm": 5.787412256354303,
"learning_rate": 1e-06,
"loss": 0.2285,
"step": 3428
},
{
"epoch": 1.3112810707456979,
"grad_norm": 3.2702141598908003,
"learning_rate": 1e-06,
"loss": 0.4435,
"step": 3429
},
{
"epoch": 1.311663479923518,
"grad_norm": 2.8847752120905352,
"learning_rate": 1e-06,
"loss": 0.1651,
"step": 3430
},
{
"epoch": 1.3120458891013385,
"grad_norm": 2.979687909674854,
"learning_rate": 1e-06,
"loss": 0.3365,
"step": 3431
},
{
"epoch": 1.3124282982791586,
"grad_norm": 3.408940363941151,
"learning_rate": 1e-06,
"loss": 0.2235,
"step": 3432
},
{
"epoch": 1.312810707456979,
"grad_norm": 2.310862709057707,
"learning_rate": 1e-06,
"loss": 0.2305,
"step": 3433
},
{
"epoch": 1.3131931166347992,
"grad_norm": 3.007345425889769,
"learning_rate": 1e-06,
"loss": 0.2149,
"step": 3434
},
{
"epoch": 1.3135755258126194,
"grad_norm": 2.668927803638142,
"learning_rate": 1e-06,
"loss": 0.3073,
"step": 3435
},
{
"epoch": 1.3139579349904398,
"grad_norm": 2.104125321488783,
"learning_rate": 1e-06,
"loss": 0.1248,
"step": 3436
},
{
"epoch": 1.31434034416826,
"grad_norm": 2.6199810548636107,
"learning_rate": 1e-06,
"loss": 0.1346,
"step": 3437
},
{
"epoch": 1.3147227533460804,
"grad_norm": 5.242819462303336,
"learning_rate": 1e-06,
"loss": 0.1127,
"step": 3438
},
{
"epoch": 1.3151051625239005,
"grad_norm": 2.5933172025726394,
"learning_rate": 1e-06,
"loss": 0.0724,
"step": 3439
},
{
"epoch": 1.3154875717017207,
"grad_norm": 3.417160339666154,
"learning_rate": 1e-06,
"loss": 0.1142,
"step": 3440
},
{
"epoch": 1.315869980879541,
"grad_norm": 5.838230729366253,
"learning_rate": 1e-06,
"loss": 0.1927,
"step": 3441
},
{
"epoch": 1.3162523900573615,
"grad_norm": 3.554357089941153,
"learning_rate": 1e-06,
"loss": 0.2933,
"step": 3442
},
{
"epoch": 1.3166347992351817,
"grad_norm": 3.5519096503580494,
"learning_rate": 1e-06,
"loss": 0.319,
"step": 3443
},
{
"epoch": 1.3170172084130018,
"grad_norm": 2.831211155504948,
"learning_rate": 1e-06,
"loss": 0.2755,
"step": 3444
},
{
"epoch": 1.3173996175908222,
"grad_norm": 3.2813162660719373,
"learning_rate": 1e-06,
"loss": 0.2042,
"step": 3445
},
{
"epoch": 1.3177820267686424,
"grad_norm": 3.371527404121634,
"learning_rate": 1e-06,
"loss": 0.2246,
"step": 3446
},
{
"epoch": 1.3181644359464628,
"grad_norm": 2.7169618535907323,
"learning_rate": 1e-06,
"loss": 0.1714,
"step": 3447
},
{
"epoch": 1.318546845124283,
"grad_norm": 2.2601980522607117,
"learning_rate": 1e-06,
"loss": 0.0768,
"step": 3448
},
{
"epoch": 1.3189292543021032,
"grad_norm": 2.448373846554622,
"learning_rate": 1e-06,
"loss": 0.1442,
"step": 3449
},
{
"epoch": 1.3193116634799236,
"grad_norm": 4.5222297070623805,
"learning_rate": 1e-06,
"loss": 0.1773,
"step": 3450
},
{
"epoch": 1.3196940726577437,
"grad_norm": 2.897869866818496,
"learning_rate": 1e-06,
"loss": 0.08,
"step": 3451
},
{
"epoch": 1.3200764818355641,
"grad_norm": 3.647040743032006,
"learning_rate": 1e-06,
"loss": 0.1009,
"step": 3452
},
{
"epoch": 1.3204588910133843,
"grad_norm": 4.816123490027161,
"learning_rate": 1e-06,
"loss": 0.1705,
"step": 3453
},
{
"epoch": 1.3208413001912045,
"grad_norm": 4.641099693603532,
"learning_rate": 1e-06,
"loss": 0.5515,
"step": 3454
},
{
"epoch": 1.321223709369025,
"grad_norm": 4.405594797321751,
"learning_rate": 1e-06,
"loss": 0.4077,
"step": 3455
},
{
"epoch": 1.321606118546845,
"grad_norm": 4.125408730342099,
"learning_rate": 1e-06,
"loss": 0.3148,
"step": 3456
},
{
"epoch": 1.3219885277246655,
"grad_norm": 3.544837938055854,
"learning_rate": 1e-06,
"loss": 0.3524,
"step": 3457
},
{
"epoch": 1.3223709369024856,
"grad_norm": 3.337091393764473,
"learning_rate": 1e-06,
"loss": 0.4117,
"step": 3458
},
{
"epoch": 1.3227533460803058,
"grad_norm": 3.2285444254897344,
"learning_rate": 1e-06,
"loss": 0.2332,
"step": 3459
},
{
"epoch": 1.3231357552581262,
"grad_norm": 3.494300697886279,
"learning_rate": 1e-06,
"loss": 0.1988,
"step": 3460
},
{
"epoch": 1.3235181644359464,
"grad_norm": 3.257444511871968,
"learning_rate": 1e-06,
"loss": 0.204,
"step": 3461
},
{
"epoch": 1.3239005736137668,
"grad_norm": 3.4655272488098623,
"learning_rate": 1e-06,
"loss": 0.2455,
"step": 3462
},
{
"epoch": 1.324282982791587,
"grad_norm": 4.786664701654426,
"learning_rate": 1e-06,
"loss": 0.1595,
"step": 3463
},
{
"epoch": 1.3246653919694071,
"grad_norm": 3.0983968958115438,
"learning_rate": 1e-06,
"loss": 0.1049,
"step": 3464
},
{
"epoch": 1.3250478011472275,
"grad_norm": 4.1737502094827414,
"learning_rate": 1e-06,
"loss": 0.0904,
"step": 3465
},
{
"epoch": 1.325430210325048,
"grad_norm": 5.634086898572302,
"learning_rate": 1e-06,
"loss": 0.5333,
"step": 3466
},
{
"epoch": 1.325812619502868,
"grad_norm": 3.880113734515193,
"learning_rate": 1e-06,
"loss": 0.524,
"step": 3467
},
{
"epoch": 1.3261950286806883,
"grad_norm": 3.870828906256228,
"learning_rate": 1e-06,
"loss": 0.3742,
"step": 3468
},
{
"epoch": 1.3265774378585087,
"grad_norm": 2.1852502425555547,
"learning_rate": 1e-06,
"loss": 0.3189,
"step": 3469
},
{
"epoch": 1.3269598470363289,
"grad_norm": 3.554307317838956,
"learning_rate": 1e-06,
"loss": 0.1516,
"step": 3470
},
{
"epoch": 1.3273422562141493,
"grad_norm": 2.980956511376388,
"learning_rate": 1e-06,
"loss": 0.2794,
"step": 3471
},
{
"epoch": 1.3277246653919694,
"grad_norm": 2.975385938983526,
"learning_rate": 1e-06,
"loss": 0.1522,
"step": 3472
},
{
"epoch": 1.3281070745697896,
"grad_norm": 3.4128798612232973,
"learning_rate": 1e-06,
"loss": 0.1684,
"step": 3473
},
{
"epoch": 1.32848948374761,
"grad_norm": 3.013378037512129,
"learning_rate": 1e-06,
"loss": 0.1744,
"step": 3474
},
{
"epoch": 1.3288718929254302,
"grad_norm": 3.5179992796445525,
"learning_rate": 1e-06,
"loss": 0.2799,
"step": 3475
},
{
"epoch": 1.3292543021032506,
"grad_norm": 3.902518970118376,
"learning_rate": 1e-06,
"loss": 0.1032,
"step": 3476
},
{
"epoch": 1.3296367112810707,
"grad_norm": 2.980508586572323,
"learning_rate": 1e-06,
"loss": 0.0781,
"step": 3477
},
{
"epoch": 1.330019120458891,
"grad_norm": 4.325377325455937,
"learning_rate": 1e-06,
"loss": 0.1956,
"step": 3478
},
{
"epoch": 1.3304015296367113,
"grad_norm": 4.729012552626647,
"learning_rate": 1e-06,
"loss": 0.4777,
"step": 3479
},
{
"epoch": 1.3307839388145315,
"grad_norm": 3.6483533306337645,
"learning_rate": 1e-06,
"loss": 0.4687,
"step": 3480
},
{
"epoch": 1.331166347992352,
"grad_norm": 2.6517313501639608,
"learning_rate": 1e-06,
"loss": 0.2736,
"step": 3481
},
{
"epoch": 1.331548757170172,
"grad_norm": 2.55513038645958,
"learning_rate": 1e-06,
"loss": 0.1905,
"step": 3482
},
{
"epoch": 1.3319311663479922,
"grad_norm": 2.823382790757958,
"learning_rate": 1e-06,
"loss": 0.1812,
"step": 3483
},
{
"epoch": 1.3323135755258126,
"grad_norm": 3.1726080888231554,
"learning_rate": 1e-06,
"loss": 0.1316,
"step": 3484
},
{
"epoch": 1.3326959847036328,
"grad_norm": 3.4896985087200147,
"learning_rate": 1e-06,
"loss": 0.2107,
"step": 3485
},
{
"epoch": 1.3330783938814532,
"grad_norm": 3.3192770386801516,
"learning_rate": 1e-06,
"loss": 0.2316,
"step": 3486
},
{
"epoch": 1.3334608030592734,
"grad_norm": 3.5117125039050405,
"learning_rate": 1e-06,
"loss": 0.1562,
"step": 3487
},
{
"epoch": 1.3338432122370936,
"grad_norm": 4.913830386260947,
"learning_rate": 1e-06,
"loss": 0.0863,
"step": 3488
},
{
"epoch": 1.334225621414914,
"grad_norm": 2.962580645143604,
"learning_rate": 1e-06,
"loss": 0.0658,
"step": 3489
},
{
"epoch": 1.3346080305927344,
"grad_norm": 4.366650051832743,
"learning_rate": 1e-06,
"loss": 0.2274,
"step": 3490
},
{
"epoch": 1.3349904397705545,
"grad_norm": 5.46336980742746,
"learning_rate": 1e-06,
"loss": 0.3706,
"step": 3491
},
{
"epoch": 1.3353728489483747,
"grad_norm": 5.000078009950527,
"learning_rate": 1e-06,
"loss": 0.2676,
"step": 3492
},
{
"epoch": 1.335755258126195,
"grad_norm": 5.653906236505976,
"learning_rate": 1e-06,
"loss": 0.2413,
"step": 3493
},
{
"epoch": 1.3361376673040153,
"grad_norm": 3.2929180679855703,
"learning_rate": 1e-06,
"loss": 0.3646,
"step": 3494
},
{
"epoch": 1.3365200764818357,
"grad_norm": 3.7268240335047995,
"learning_rate": 1e-06,
"loss": 0.5141,
"step": 3495
},
{
"epoch": 1.3369024856596559,
"grad_norm": 3.816759575347373,
"learning_rate": 1e-06,
"loss": 0.225,
"step": 3496
},
{
"epoch": 1.337284894837476,
"grad_norm": 2.923530476823112,
"learning_rate": 1e-06,
"loss": 0.2098,
"step": 3497
},
{
"epoch": 1.3376673040152964,
"grad_norm": 2.427082984062028,
"learning_rate": 1e-06,
"loss": 0.1571,
"step": 3498
},
{
"epoch": 1.3380497131931166,
"grad_norm": 2.846076935543849,
"learning_rate": 1e-06,
"loss": 0.1271,
"step": 3499
},
{
"epoch": 1.338432122370937,
"grad_norm": 2.8243415774153005,
"learning_rate": 1e-06,
"loss": 0.1141,
"step": 3500
},
{
"epoch": 1.338432122370937,
"eval_runtime": 829.2431,
"eval_samples_per_second": 1.85,
"eval_steps_per_second": 0.463,
"step": 3500
},
{
"epoch": 1.3388145315487572,
"grad_norm": 3.799385904584203,
"learning_rate": 1e-06,
"loss": 0.1167,
"step": 3501
},
{
"epoch": 1.3391969407265774,
"grad_norm": 3.888971919732548,
"learning_rate": 1e-06,
"loss": 0.0944,
"step": 3502
},
{
"epoch": 1.3395793499043978,
"grad_norm": 3.5823894558746696,
"learning_rate": 1e-06,
"loss": 0.2139,
"step": 3503
},
{
"epoch": 1.339961759082218,
"grad_norm": 4.06310021294751,
"learning_rate": 1e-06,
"loss": 0.2905,
"step": 3504
},
{
"epoch": 1.3403441682600383,
"grad_norm": 3.404893237441377,
"learning_rate": 1e-06,
"loss": 0.2061,
"step": 3505
},
{
"epoch": 1.3407265774378585,
"grad_norm": 2.0965391342036903,
"learning_rate": 1e-06,
"loss": 0.286,
"step": 3506
},
{
"epoch": 1.3411089866156787,
"grad_norm": 2.977220037127773,
"learning_rate": 1e-06,
"loss": 0.1368,
"step": 3507
},
{
"epoch": 1.341491395793499,
"grad_norm": 3.686750174111812,
"learning_rate": 1e-06,
"loss": 0.345,
"step": 3508
},
{
"epoch": 1.3418738049713193,
"grad_norm": 3.084022565222293,
"learning_rate": 1e-06,
"loss": 0.2389,
"step": 3509
},
{
"epoch": 1.3422562141491396,
"grad_norm": 3.6780032699263425,
"learning_rate": 1e-06,
"loss": 0.2237,
"step": 3510
},
{
"epoch": 1.3426386233269598,
"grad_norm": 3.4518122184220217,
"learning_rate": 1e-06,
"loss": 0.1802,
"step": 3511
},
{
"epoch": 1.34302103250478,
"grad_norm": 3.9095442652174097,
"learning_rate": 1e-06,
"loss": 0.169,
"step": 3512
},
{
"epoch": 1.3434034416826004,
"grad_norm": 3.3225856763644526,
"learning_rate": 1e-06,
"loss": 0.0833,
"step": 3513
},
{
"epoch": 1.3437858508604206,
"grad_norm": 4.5201920905692194,
"learning_rate": 1e-06,
"loss": 0.0814,
"step": 3514
},
{
"epoch": 1.344168260038241,
"grad_norm": 4.5273921986999985,
"learning_rate": 1e-06,
"loss": 0.1546,
"step": 3515
},
{
"epoch": 1.3445506692160611,
"grad_norm": 5.1095078943917445,
"learning_rate": 1e-06,
"loss": 0.2919,
"step": 3516
},
{
"epoch": 1.3449330783938813,
"grad_norm": 3.249637730288223,
"learning_rate": 1e-06,
"loss": 0.5484,
"step": 3517
},
{
"epoch": 1.3453154875717017,
"grad_norm": 2.9656931347752957,
"learning_rate": 1e-06,
"loss": 0.269,
"step": 3518
},
{
"epoch": 1.3456978967495221,
"grad_norm": 3.563711127822158,
"learning_rate": 1e-06,
"loss": 0.1455,
"step": 3519
},
{
"epoch": 1.3460803059273423,
"grad_norm": 2.7947434129126334,
"learning_rate": 1e-06,
"loss": 0.328,
"step": 3520
},
{
"epoch": 1.3464627151051625,
"grad_norm": 2.752041145904254,
"learning_rate": 1e-06,
"loss": 0.2037,
"step": 3521
},
{
"epoch": 1.3468451242829829,
"grad_norm": 3.3458457867369633,
"learning_rate": 1e-06,
"loss": 0.1078,
"step": 3522
},
{
"epoch": 1.347227533460803,
"grad_norm": 2.7651796143950076,
"learning_rate": 1e-06,
"loss": 0.1771,
"step": 3523
},
{
"epoch": 1.3476099426386234,
"grad_norm": 3.3446008188042406,
"learning_rate": 1e-06,
"loss": 0.1971,
"step": 3524
},
{
"epoch": 1.3479923518164436,
"grad_norm": 3.924573479092709,
"learning_rate": 1e-06,
"loss": 0.2425,
"step": 3525
},
{
"epoch": 1.3483747609942638,
"grad_norm": 4.163756740430623,
"learning_rate": 1e-06,
"loss": 0.1407,
"step": 3526
},
{
"epoch": 1.3487571701720842,
"grad_norm": 2.5869219885330637,
"learning_rate": 1e-06,
"loss": 0.057,
"step": 3527
},
{
"epoch": 1.3491395793499044,
"grad_norm": 6.614779379835467,
"learning_rate": 1e-06,
"loss": 0.2318,
"step": 3528
},
{
"epoch": 1.3495219885277248,
"grad_norm": 6.1811222383353766,
"learning_rate": 1e-06,
"loss": 0.4225,
"step": 3529
},
{
"epoch": 1.349904397705545,
"grad_norm": 3.8939866093589983,
"learning_rate": 1e-06,
"loss": 0.53,
"step": 3530
},
{
"epoch": 1.350286806883365,
"grad_norm": 2.486690857371599,
"learning_rate": 1e-06,
"loss": 0.3089,
"step": 3531
},
{
"epoch": 1.3506692160611855,
"grad_norm": 3.0648838029899403,
"learning_rate": 1e-06,
"loss": 0.337,
"step": 3532
},
{
"epoch": 1.3510516252390057,
"grad_norm": 4.054914699649741,
"learning_rate": 1e-06,
"loss": 0.4405,
"step": 3533
},
{
"epoch": 1.351434034416826,
"grad_norm": 2.605171060784863,
"learning_rate": 1e-06,
"loss": 0.1541,
"step": 3534
},
{
"epoch": 1.3518164435946463,
"grad_norm": 2.86937046363919,
"learning_rate": 1e-06,
"loss": 0.1522,
"step": 3535
},
{
"epoch": 1.3521988527724664,
"grad_norm": 3.984496829096379,
"learning_rate": 1e-06,
"loss": 0.2619,
"step": 3536
},
{
"epoch": 1.3525812619502868,
"grad_norm": 3.5482169818093143,
"learning_rate": 1e-06,
"loss": 0.2178,
"step": 3537
},
{
"epoch": 1.352963671128107,
"grad_norm": 2.611337985560104,
"learning_rate": 1e-06,
"loss": 0.0677,
"step": 3538
},
{
"epoch": 1.3533460803059274,
"grad_norm": 3.1567322013306685,
"learning_rate": 1e-06,
"loss": 0.0957,
"step": 3539
},
{
"epoch": 1.3537284894837476,
"grad_norm": 8.004134540271641,
"learning_rate": 1e-06,
"loss": 0.1923,
"step": 3540
},
{
"epoch": 1.3541108986615678,
"grad_norm": 4.335753620819479,
"learning_rate": 1e-06,
"loss": 0.3444,
"step": 3541
},
{
"epoch": 1.3544933078393881,
"grad_norm": 3.0156895131538484,
"learning_rate": 1e-06,
"loss": 0.3191,
"step": 3542
},
{
"epoch": 1.3548757170172085,
"grad_norm": 3.65331430850215,
"learning_rate": 1e-06,
"loss": 0.2218,
"step": 3543
},
{
"epoch": 1.3552581261950287,
"grad_norm": 2.125882806844032,
"learning_rate": 1e-06,
"loss": 0.1969,
"step": 3544
},
{
"epoch": 1.355640535372849,
"grad_norm": 3.1973091733060524,
"learning_rate": 1e-06,
"loss": 0.3326,
"step": 3545
},
{
"epoch": 1.3560229445506693,
"grad_norm": 3.9776660159966237,
"learning_rate": 1e-06,
"loss": 0.3211,
"step": 3546
},
{
"epoch": 1.3564053537284895,
"grad_norm": 3.0101871776996854,
"learning_rate": 1e-06,
"loss": 0.2689,
"step": 3547
},
{
"epoch": 1.3567877629063099,
"grad_norm": 4.576216889183216,
"learning_rate": 1e-06,
"loss": 0.1728,
"step": 3548
},
{
"epoch": 1.35717017208413,
"grad_norm": 3.005881108037498,
"learning_rate": 1e-06,
"loss": 0.1002,
"step": 3549
},
{
"epoch": 1.3575525812619502,
"grad_norm": 3.1832495380444117,
"learning_rate": 1e-06,
"loss": 0.129,
"step": 3550
},
{
"epoch": 1.3579349904397706,
"grad_norm": 3.0458417656638703,
"learning_rate": 1e-06,
"loss": 0.1105,
"step": 3551
},
{
"epoch": 1.3583173996175908,
"grad_norm": 3.0338615886722966,
"learning_rate": 1e-06,
"loss": 0.0667,
"step": 3552
},
{
"epoch": 1.3586998087954112,
"grad_norm": 4.363334798868475,
"learning_rate": 1e-06,
"loss": 0.4191,
"step": 3553
},
{
"epoch": 1.3590822179732314,
"grad_norm": 3.7229429056573213,
"learning_rate": 1e-06,
"loss": 0.2664,
"step": 3554
},
{
"epoch": 1.3594646271510515,
"grad_norm": 2.870013016727716,
"learning_rate": 1e-06,
"loss": 0.3336,
"step": 3555
},
{
"epoch": 1.359847036328872,
"grad_norm": 2.168764700276357,
"learning_rate": 1e-06,
"loss": 0.1946,
"step": 3556
},
{
"epoch": 1.3602294455066921,
"grad_norm": 3.44598478993819,
"learning_rate": 1e-06,
"loss": 0.2641,
"step": 3557
},
{
"epoch": 1.3606118546845125,
"grad_norm": 4.045256655629172,
"learning_rate": 1e-06,
"loss": 0.4309,
"step": 3558
},
{
"epoch": 1.3609942638623327,
"grad_norm": 3.2178948016281863,
"learning_rate": 1e-06,
"loss": 0.3445,
"step": 3559
},
{
"epoch": 1.3613766730401529,
"grad_norm": 2.952558836580853,
"learning_rate": 1e-06,
"loss": 0.2832,
"step": 3560
},
{
"epoch": 1.3617590822179733,
"grad_norm": 2.876979643886604,
"learning_rate": 1e-06,
"loss": 0.209,
"step": 3561
},
{
"epoch": 1.3621414913957934,
"grad_norm": 3.4246323158433025,
"learning_rate": 1e-06,
"loss": 0.0917,
"step": 3562
},
{
"epoch": 1.3625239005736138,
"grad_norm": 3.293764500160801,
"learning_rate": 1e-06,
"loss": 0.0904,
"step": 3563
},
{
"epoch": 1.362906309751434,
"grad_norm": 3.631939526747678,
"learning_rate": 1e-06,
"loss": 0.0889,
"step": 3564
},
{
"epoch": 1.3632887189292542,
"grad_norm": 4.249874786327579,
"learning_rate": 1e-06,
"loss": 0.0662,
"step": 3565
},
{
"epoch": 1.3636711281070746,
"grad_norm": 8.425726468151769,
"learning_rate": 1e-06,
"loss": 0.2319,
"step": 3566
},
{
"epoch": 1.3640535372848948,
"grad_norm": 4.157774681426793,
"learning_rate": 1e-06,
"loss": 0.4404,
"step": 3567
},
{
"epoch": 1.3644359464627152,
"grad_norm": 3.2119620143596737,
"learning_rate": 1e-06,
"loss": 0.2398,
"step": 3568
},
{
"epoch": 1.3648183556405353,
"grad_norm": 2.3561767728643264,
"learning_rate": 1e-06,
"loss": 0.2127,
"step": 3569
},
{
"epoch": 1.3652007648183555,
"grad_norm": 3.921888860548839,
"learning_rate": 1e-06,
"loss": 0.5913,
"step": 3570
},
{
"epoch": 1.365583173996176,
"grad_norm": 3.304847659294973,
"learning_rate": 1e-06,
"loss": 0.1353,
"step": 3571
},
{
"epoch": 1.3659655831739963,
"grad_norm": 3.5156186252112342,
"learning_rate": 1e-06,
"loss": 0.1858,
"step": 3572
},
{
"epoch": 1.3663479923518165,
"grad_norm": 2.9621940100661,
"learning_rate": 1e-06,
"loss": 0.1808,
"step": 3573
},
{
"epoch": 1.3667304015296367,
"grad_norm": 3.2755091264155327,
"learning_rate": 1e-06,
"loss": 0.2464,
"step": 3574
},
{
"epoch": 1.367112810707457,
"grad_norm": 4.7476897393709425,
"learning_rate": 1e-06,
"loss": 0.297,
"step": 3575
},
{
"epoch": 1.3674952198852772,
"grad_norm": 3.513067781979689,
"learning_rate": 1e-06,
"loss": 0.1809,
"step": 3576
},
{
"epoch": 1.3678776290630976,
"grad_norm": 3.219846816358348,
"learning_rate": 1e-06,
"loss": 0.0732,
"step": 3577
},
{
"epoch": 1.3682600382409178,
"grad_norm": 6.085868864443686,
"learning_rate": 1e-06,
"loss": 0.1976,
"step": 3578
},
{
"epoch": 1.368642447418738,
"grad_norm": 4.61869093014494,
"learning_rate": 1e-06,
"loss": 0.6497,
"step": 3579
},
{
"epoch": 1.3690248565965584,
"grad_norm": 4.839491023895109,
"learning_rate": 1e-06,
"loss": 0.3992,
"step": 3580
},
{
"epoch": 1.3694072657743785,
"grad_norm": 2.65189525200031,
"learning_rate": 1e-06,
"loss": 0.2308,
"step": 3581
},
{
"epoch": 1.369789674952199,
"grad_norm": 4.629822562259575,
"learning_rate": 1e-06,
"loss": 0.3704,
"step": 3582
},
{
"epoch": 1.3701720841300191,
"grad_norm": 4.7582316096578845,
"learning_rate": 1e-06,
"loss": 0.4397,
"step": 3583
},
{
"epoch": 1.3705544933078393,
"grad_norm": 2.7484996344052743,
"learning_rate": 1e-06,
"loss": 0.174,
"step": 3584
},
{
"epoch": 1.3709369024856597,
"grad_norm": 4.355188909872203,
"learning_rate": 1e-06,
"loss": 0.1491,
"step": 3585
},
{
"epoch": 1.3713193116634799,
"grad_norm": 3.2913755135195517,
"learning_rate": 1e-06,
"loss": 0.1138,
"step": 3586
},
{
"epoch": 1.3717017208413003,
"grad_norm": 3.575477920278412,
"learning_rate": 1e-06,
"loss": 0.2969,
"step": 3587
},
{
"epoch": 1.3720841300191204,
"grad_norm": 3.3627597524155393,
"learning_rate": 1e-06,
"loss": 0.1533,
"step": 3588
},
{
"epoch": 1.3724665391969406,
"grad_norm": 3.8016096821658465,
"learning_rate": 1e-06,
"loss": 0.0731,
"step": 3589
},
{
"epoch": 1.372848948374761,
"grad_norm": 7.239808911072699,
"learning_rate": 1e-06,
"loss": 0.2692,
"step": 3590
},
{
"epoch": 1.3732313575525812,
"grad_norm": 7.992433784695021,
"learning_rate": 1e-06,
"loss": 0.5004,
"step": 3591
},
{
"epoch": 1.3736137667304016,
"grad_norm": 2.7557299480149937,
"learning_rate": 1e-06,
"loss": 0.3263,
"step": 3592
},
{
"epoch": 1.3739961759082218,
"grad_norm": 3.175777496298135,
"learning_rate": 1e-06,
"loss": 0.2712,
"step": 3593
},
{
"epoch": 1.374378585086042,
"grad_norm": 2.3902367170571446,
"learning_rate": 1e-06,
"loss": 0.3309,
"step": 3594
},
{
"epoch": 1.3747609942638623,
"grad_norm": 2.9567911310100063,
"learning_rate": 1e-06,
"loss": 0.2152,
"step": 3595
},
{
"epoch": 1.3751434034416827,
"grad_norm": 3.233958931666279,
"learning_rate": 1e-06,
"loss": 0.3663,
"step": 3596
},
{
"epoch": 1.375525812619503,
"grad_norm": 3.6868141959014014,
"learning_rate": 1e-06,
"loss": 0.2287,
"step": 3597
},
{
"epoch": 1.375908221797323,
"grad_norm": 2.6929733962463547,
"learning_rate": 1e-06,
"loss": 0.1352,
"step": 3598
},
{
"epoch": 1.3762906309751435,
"grad_norm": 3.3579409355072016,
"learning_rate": 1e-06,
"loss": 0.1688,
"step": 3599
},
{
"epoch": 1.3766730401529637,
"grad_norm": 4.225536623852819,
"learning_rate": 1e-06,
"loss": 0.1663,
"step": 3600
},
{
"epoch": 1.377055449330784,
"grad_norm": 6.74606886938087,
"learning_rate": 1e-06,
"loss": 0.3401,
"step": 3601
},
{
"epoch": 1.3774378585086042,
"grad_norm": 3.743459592837295,
"learning_rate": 1e-06,
"loss": 0.0814,
"step": 3602
},
{
"epoch": 1.3778202676864244,
"grad_norm": 9.690145414381327,
"learning_rate": 1e-06,
"loss": 0.2664,
"step": 3603
},
{
"epoch": 1.3782026768642448,
"grad_norm": 3.8949521678098136,
"learning_rate": 1e-06,
"loss": 0.3333,
"step": 3604
},
{
"epoch": 1.378585086042065,
"grad_norm": 3.4412114307973285,
"learning_rate": 1e-06,
"loss": 0.3546,
"step": 3605
},
{
"epoch": 1.3789674952198854,
"grad_norm": 2.817739226805909,
"learning_rate": 1e-06,
"loss": 0.3284,
"step": 3606
},
{
"epoch": 1.3793499043977056,
"grad_norm": 4.063424342708501,
"learning_rate": 1e-06,
"loss": 0.3598,
"step": 3607
},
{
"epoch": 1.3797323135755257,
"grad_norm": 2.7788390950734434,
"learning_rate": 1e-06,
"loss": 0.1863,
"step": 3608
},
{
"epoch": 1.3801147227533461,
"grad_norm": 2.5297091468417365,
"learning_rate": 1e-06,
"loss": 0.1796,
"step": 3609
},
{
"epoch": 1.3804971319311663,
"grad_norm": 2.5535086148641253,
"learning_rate": 1e-06,
"loss": 0.151,
"step": 3610
},
{
"epoch": 1.3808795411089867,
"grad_norm": 4.293329503154428,
"learning_rate": 1e-06,
"loss": 0.381,
"step": 3611
},
{
"epoch": 1.3812619502868069,
"grad_norm": 2.824774343204578,
"learning_rate": 1e-06,
"loss": 0.1622,
"step": 3612
},
{
"epoch": 1.381644359464627,
"grad_norm": 2.6020886688136997,
"learning_rate": 1e-06,
"loss": 0.073,
"step": 3613
},
{
"epoch": 1.3820267686424474,
"grad_norm": 2.7319932428279214,
"learning_rate": 1e-06,
"loss": 0.0816,
"step": 3614
},
{
"epoch": 1.3824091778202676,
"grad_norm": 3.8276916258599285,
"learning_rate": 1e-06,
"loss": 0.1063,
"step": 3615
},
{
"epoch": 1.382791586998088,
"grad_norm": 5.450492332688963,
"learning_rate": 1e-06,
"loss": 0.4124,
"step": 3616
},
{
"epoch": 1.3831739961759082,
"grad_norm": 4.239795776343202,
"learning_rate": 1e-06,
"loss": 0.5117,
"step": 3617
},
{
"epoch": 1.3835564053537284,
"grad_norm": 3.7588389178542854,
"learning_rate": 1e-06,
"loss": 0.3095,
"step": 3618
},
{
"epoch": 1.3839388145315488,
"grad_norm": 3.728928172975965,
"learning_rate": 1e-06,
"loss": 0.4486,
"step": 3619
},
{
"epoch": 1.384321223709369,
"grad_norm": 5.153491952946277,
"learning_rate": 1e-06,
"loss": 0.2694,
"step": 3620
},
{
"epoch": 1.3847036328871893,
"grad_norm": 3.255781167045849,
"learning_rate": 1e-06,
"loss": 0.2452,
"step": 3621
},
{
"epoch": 1.3850860420650095,
"grad_norm": 2.844236059883789,
"learning_rate": 1e-06,
"loss": 0.2148,
"step": 3622
},
{
"epoch": 1.3854684512428297,
"grad_norm": 2.263122439252642,
"learning_rate": 1e-06,
"loss": 0.2454,
"step": 3623
},
{
"epoch": 1.38585086042065,
"grad_norm": 2.4557283036738498,
"learning_rate": 1e-06,
"loss": 0.1102,
"step": 3624
},
{
"epoch": 1.3862332695984705,
"grad_norm": 3.1210013174184397,
"learning_rate": 1e-06,
"loss": 0.0913,
"step": 3625
},
{
"epoch": 1.3866156787762907,
"grad_norm": 2.220231756880406,
"learning_rate": 1e-06,
"loss": 0.0711,
"step": 3626
},
{
"epoch": 1.3869980879541108,
"grad_norm": 3.0206513097885277,
"learning_rate": 1e-06,
"loss": 0.106,
"step": 3627
},
{
"epoch": 1.3873804971319312,
"grad_norm": 3.5395434728529356,
"learning_rate": 1e-06,
"loss": 0.2138,
"step": 3628
},
{
"epoch": 1.3877629063097514,
"grad_norm": 14.14533653379434,
"learning_rate": 1e-06,
"loss": 0.3418,
"step": 3629
},
{
"epoch": 1.3881453154875718,
"grad_norm": 3.6628960092640654,
"learning_rate": 1e-06,
"loss": 0.5699,
"step": 3630
},
{
"epoch": 1.388527724665392,
"grad_norm": 2.6896311272339815,
"learning_rate": 1e-06,
"loss": 0.208,
"step": 3631
},
{
"epoch": 1.3889101338432122,
"grad_norm": 3.044524709505099,
"learning_rate": 1e-06,
"loss": 0.2483,
"step": 3632
},
{
"epoch": 1.3892925430210326,
"grad_norm": 3.045121377865358,
"learning_rate": 1e-06,
"loss": 0.2137,
"step": 3633
},
{
"epoch": 1.3896749521988527,
"grad_norm": 3.8082912540128318,
"learning_rate": 1e-06,
"loss": 0.314,
"step": 3634
},
{
"epoch": 1.3900573613766731,
"grad_norm": 2.786056052192463,
"learning_rate": 1e-06,
"loss": 0.202,
"step": 3635
},
{
"epoch": 1.3904397705544933,
"grad_norm": 2.5837768563626664,
"learning_rate": 1e-06,
"loss": 0.2055,
"step": 3636
},
{
"epoch": 1.3908221797323135,
"grad_norm": 3.4122340283827435,
"learning_rate": 1e-06,
"loss": 0.2463,
"step": 3637
},
{
"epoch": 1.3912045889101339,
"grad_norm": 4.804509254188609,
"learning_rate": 1e-06,
"loss": 0.1585,
"step": 3638
},
{
"epoch": 1.391586998087954,
"grad_norm": 3.546749700946221,
"learning_rate": 1e-06,
"loss": 0.0743,
"step": 3639
},
{
"epoch": 1.3919694072657744,
"grad_norm": 3.133098788985979,
"learning_rate": 1e-06,
"loss": 0.0622,
"step": 3640
},
{
"epoch": 1.3923518164435946,
"grad_norm": 6.5644693961788825,
"learning_rate": 1e-06,
"loss": 0.281,
"step": 3641
},
{
"epoch": 1.3927342256214148,
"grad_norm": 3.2124314000267664,
"learning_rate": 1e-06,
"loss": 0.3506,
"step": 3642
},
{
"epoch": 1.3931166347992352,
"grad_norm": 2.6462614859438487,
"learning_rate": 1e-06,
"loss": 0.3129,
"step": 3643
},
{
"epoch": 1.3934990439770554,
"grad_norm": 2.184482018464643,
"learning_rate": 1e-06,
"loss": 0.1818,
"step": 3644
},
{
"epoch": 1.3938814531548758,
"grad_norm": 2.879734122136413,
"learning_rate": 1e-06,
"loss": 0.2177,
"step": 3645
},
{
"epoch": 1.394263862332696,
"grad_norm": 2.2983385470605278,
"learning_rate": 1e-06,
"loss": 0.2533,
"step": 3646
},
{
"epoch": 1.3946462715105161,
"grad_norm": 2.7621492430731402,
"learning_rate": 1e-06,
"loss": 0.1919,
"step": 3647
},
{
"epoch": 1.3950286806883365,
"grad_norm": 3.2641662551948443,
"learning_rate": 1e-06,
"loss": 0.2589,
"step": 3648
},
{
"epoch": 1.395411089866157,
"grad_norm": 3.2110880795554464,
"learning_rate": 1e-06,
"loss": 0.1783,
"step": 3649
},
{
"epoch": 1.395793499043977,
"grad_norm": 4.443205483255379,
"learning_rate": 1e-06,
"loss": 0.1439,
"step": 3650
},
{
"epoch": 1.3961759082217973,
"grad_norm": 3.011841609442816,
"learning_rate": 1e-06,
"loss": 0.2,
"step": 3651
},
{
"epoch": 1.3965583173996177,
"grad_norm": 3.9484192353019756,
"learning_rate": 1e-06,
"loss": 0.1176,
"step": 3652
},
{
"epoch": 1.3969407265774378,
"grad_norm": 4.166531115552269,
"learning_rate": 1e-06,
"loss": 0.1474,
"step": 3653
},
{
"epoch": 1.3973231357552582,
"grad_norm": 5.26524301835712,
"learning_rate": 1e-06,
"loss": 0.5814,
"step": 3654
},
{
"epoch": 1.3977055449330784,
"grad_norm": 3.057540926717856,
"learning_rate": 1e-06,
"loss": 0.2639,
"step": 3655
},
{
"epoch": 1.3980879541108986,
"grad_norm": 2.015501269897273,
"learning_rate": 1e-06,
"loss": 0.1526,
"step": 3656
},
{
"epoch": 1.398470363288719,
"grad_norm": 2.9311632025582846,
"learning_rate": 1e-06,
"loss": 0.2517,
"step": 3657
},
{
"epoch": 1.3988527724665392,
"grad_norm": 2.72936683871947,
"learning_rate": 1e-06,
"loss": 0.1346,
"step": 3658
},
{
"epoch": 1.3992351816443596,
"grad_norm": 2.7550209332621054,
"learning_rate": 1e-06,
"loss": 0.1167,
"step": 3659
},
{
"epoch": 1.3996175908221797,
"grad_norm": 3.221825916941309,
"learning_rate": 1e-06,
"loss": 0.1641,
"step": 3660
},
{
"epoch": 1.4,
"grad_norm": 4.5738405353085545,
"learning_rate": 1e-06,
"loss": 0.2635,
"step": 3661
},
{
"epoch": 1.4003824091778203,
"grad_norm": 3.6624015508445966,
"learning_rate": 1e-06,
"loss": 0.1371,
"step": 3662
},
{
"epoch": 1.4007648183556405,
"grad_norm": 5.230158231832464,
"learning_rate": 1e-06,
"loss": 0.1611,
"step": 3663
},
{
"epoch": 1.4011472275334609,
"grad_norm": 2.4369477844717853,
"learning_rate": 1e-06,
"loss": 0.0432,
"step": 3664
},
{
"epoch": 1.401529636711281,
"grad_norm": 4.154970818335636,
"learning_rate": 1e-06,
"loss": 0.1246,
"step": 3665
},
{
"epoch": 1.4019120458891012,
"grad_norm": 6.541607834213285,
"learning_rate": 1e-06,
"loss": 0.6472,
"step": 3666
},
{
"epoch": 1.4022944550669216,
"grad_norm": 4.861430527622743,
"learning_rate": 1e-06,
"loss": 0.6617,
"step": 3667
},
{
"epoch": 1.4026768642447418,
"grad_norm": 2.6870531997099145,
"learning_rate": 1e-06,
"loss": 0.1698,
"step": 3668
},
{
"epoch": 1.4030592734225622,
"grad_norm": 2.079510105285526,
"learning_rate": 1e-06,
"loss": 0.2955,
"step": 3669
},
{
"epoch": 1.4034416826003824,
"grad_norm": 2.435760953621528,
"learning_rate": 1e-06,
"loss": 0.3218,
"step": 3670
},
{
"epoch": 1.4038240917782026,
"grad_norm": 4.586588356842911,
"learning_rate": 1e-06,
"loss": 0.2798,
"step": 3671
},
{
"epoch": 1.404206500956023,
"grad_norm": 3.088087969559685,
"learning_rate": 1e-06,
"loss": 0.2636,
"step": 3672
},
{
"epoch": 1.4045889101338433,
"grad_norm": 4.336824233652288,
"learning_rate": 1e-06,
"loss": 0.3802,
"step": 3673
},
{
"epoch": 1.4049713193116635,
"grad_norm": 3.0738729939280662,
"learning_rate": 1e-06,
"loss": 0.2326,
"step": 3674
},
{
"epoch": 1.4053537284894837,
"grad_norm": 3.5809442998820047,
"learning_rate": 1e-06,
"loss": 0.1055,
"step": 3675
},
{
"epoch": 1.405736137667304,
"grad_norm": 3.4578201342126524,
"learning_rate": 1e-06,
"loss": 0.1185,
"step": 3676
},
{
"epoch": 1.4061185468451243,
"grad_norm": 3.1254081459541228,
"learning_rate": 1e-06,
"loss": 0.1281,
"step": 3677
},
{
"epoch": 1.4065009560229447,
"grad_norm": 6.7148551119740985,
"learning_rate": 1e-06,
"loss": 0.3116,
"step": 3678
},
{
"epoch": 1.4068833652007648,
"grad_norm": 4.238429981909322,
"learning_rate": 1e-06,
"loss": 0.4395,
"step": 3679
},
{
"epoch": 1.407265774378585,
"grad_norm": 3.5441077103963172,
"learning_rate": 1e-06,
"loss": 0.2503,
"step": 3680
},
{
"epoch": 1.4076481835564054,
"grad_norm": 3.215104788916053,
"learning_rate": 1e-06,
"loss": 0.2262,
"step": 3681
},
{
"epoch": 1.4080305927342256,
"grad_norm": 4.2290451458289,
"learning_rate": 1e-06,
"loss": 0.2488,
"step": 3682
},
{
"epoch": 1.408413001912046,
"grad_norm": 3.231572065406559,
"learning_rate": 1e-06,
"loss": 0.1715,
"step": 3683
},
{
"epoch": 1.4087954110898662,
"grad_norm": 2.7355573441260024,
"learning_rate": 1e-06,
"loss": 0.2283,
"step": 3684
},
{
"epoch": 1.4091778202676863,
"grad_norm": 2.783879098283958,
"learning_rate": 1e-06,
"loss": 0.1159,
"step": 3685
},
{
"epoch": 1.4095602294455067,
"grad_norm": 3.3016736150440917,
"learning_rate": 1e-06,
"loss": 0.2454,
"step": 3686
},
{
"epoch": 1.409942638623327,
"grad_norm": 4.556744851867343,
"learning_rate": 1e-06,
"loss": 0.2448,
"step": 3687
},
{
"epoch": 1.4103250478011473,
"grad_norm": 3.576157207865198,
"learning_rate": 1e-06,
"loss": 0.1276,
"step": 3688
},
{
"epoch": 1.4107074569789675,
"grad_norm": 4.030599615579349,
"learning_rate": 1e-06,
"loss": 0.1065,
"step": 3689
},
{
"epoch": 1.4110898661567877,
"grad_norm": 3.85257161444581,
"learning_rate": 1e-06,
"loss": 0.1013,
"step": 3690
},
{
"epoch": 1.411472275334608,
"grad_norm": 6.152233691971946,
"learning_rate": 1e-06,
"loss": 0.3113,
"step": 3691
},
{
"epoch": 1.4118546845124282,
"grad_norm": 4.592085763921228,
"learning_rate": 1e-06,
"loss": 0.4697,
"step": 3692
},
{
"epoch": 1.4122370936902486,
"grad_norm": 4.530306171908304,
"learning_rate": 1e-06,
"loss": 0.3802,
"step": 3693
},
{
"epoch": 1.4126195028680688,
"grad_norm": 3.201830048596461,
"learning_rate": 1e-06,
"loss": 0.2451,
"step": 3694
},
{
"epoch": 1.413001912045889,
"grad_norm": 2.6634717717036365,
"learning_rate": 1e-06,
"loss": 0.1851,
"step": 3695
},
{
"epoch": 1.4133843212237094,
"grad_norm": 3.610814851851053,
"learning_rate": 1e-06,
"loss": 0.3227,
"step": 3696
},
{
"epoch": 1.4137667304015296,
"grad_norm": 2.707512002286083,
"learning_rate": 1e-06,
"loss": 0.2215,
"step": 3697
},
{
"epoch": 1.41414913957935,
"grad_norm": 2.7175176173168127,
"learning_rate": 1e-06,
"loss": 0.2505,
"step": 3698
},
{
"epoch": 1.4145315487571701,
"grad_norm": 3.070713628526035,
"learning_rate": 1e-06,
"loss": 0.2042,
"step": 3699
},
{
"epoch": 1.4149139579349903,
"grad_norm": 5.176987256310562,
"learning_rate": 1e-06,
"loss": 0.1853,
"step": 3700
},
{
"epoch": 1.4152963671128107,
"grad_norm": 3.5168723373172512,
"learning_rate": 1e-06,
"loss": 0.1126,
"step": 3701
},
{
"epoch": 1.415678776290631,
"grad_norm": 2.6169160161198914,
"learning_rate": 1e-06,
"loss": 0.0843,
"step": 3702
},
{
"epoch": 1.4160611854684513,
"grad_norm": 9.725306614452688,
"learning_rate": 1e-06,
"loss": 0.1681,
"step": 3703
},
{
"epoch": 1.4164435946462715,
"grad_norm": 5.82730579499387,
"learning_rate": 1e-06,
"loss": 0.5179,
"step": 3704
},
{
"epoch": 1.4168260038240919,
"grad_norm": 3.8021214034730724,
"learning_rate": 1e-06,
"loss": 0.4602,
"step": 3705
},
{
"epoch": 1.417208413001912,
"grad_norm": 3.5419228554611526,
"learning_rate": 1e-06,
"loss": 0.2668,
"step": 3706
},
{
"epoch": 1.4175908221797324,
"grad_norm": 2.426808506764268,
"learning_rate": 1e-06,
"loss": 0.1892,
"step": 3707
},
{
"epoch": 1.4179732313575526,
"grad_norm": 3.6122594099875975,
"learning_rate": 1e-06,
"loss": 0.2188,
"step": 3708
},
{
"epoch": 1.4183556405353728,
"grad_norm": 3.1404725602385586,
"learning_rate": 1e-06,
"loss": 0.3341,
"step": 3709
},
{
"epoch": 1.4187380497131932,
"grad_norm": 4.374365406333801,
"learning_rate": 1e-06,
"loss": 0.0922,
"step": 3710
},
{
"epoch": 1.4191204588910133,
"grad_norm": 3.265568344865903,
"learning_rate": 1e-06,
"loss": 0.1587,
"step": 3711
},
{
"epoch": 1.4195028680688337,
"grad_norm": 4.36933379413446,
"learning_rate": 1e-06,
"loss": 0.1202,
"step": 3712
},
{
"epoch": 1.419885277246654,
"grad_norm": 8.529275183522337,
"learning_rate": 1e-06,
"loss": 0.1127,
"step": 3713
},
{
"epoch": 1.420267686424474,
"grad_norm": 4.165621639173087,
"learning_rate": 1e-06,
"loss": 0.1621,
"step": 3714
},
{
"epoch": 1.4206500956022945,
"grad_norm": 3.330562361695334,
"learning_rate": 1e-06,
"loss": 0.1021,
"step": 3715
},
{
"epoch": 1.4210325047801147,
"grad_norm": 7.72024583415231,
"learning_rate": 1e-06,
"loss": 0.3048,
"step": 3716
},
{
"epoch": 1.421414913957935,
"grad_norm": 3.407402753442851,
"learning_rate": 1e-06,
"loss": 0.3945,
"step": 3717
},
{
"epoch": 1.4217973231357552,
"grad_norm": 3.9287353332808927,
"learning_rate": 1e-06,
"loss": 0.2428,
"step": 3718
},
{
"epoch": 1.4221797323135754,
"grad_norm": 3.744834903189784,
"learning_rate": 1e-06,
"loss": 0.1774,
"step": 3719
},
{
"epoch": 1.4225621414913958,
"grad_norm": 3.6131709231467375,
"learning_rate": 1e-06,
"loss": 0.2858,
"step": 3720
},
{
"epoch": 1.422944550669216,
"grad_norm": 3.777359403244454,
"learning_rate": 1e-06,
"loss": 0.3237,
"step": 3721
},
{
"epoch": 1.4233269598470364,
"grad_norm": 2.544264127995152,
"learning_rate": 1e-06,
"loss": 0.1212,
"step": 3722
},
{
"epoch": 1.4237093690248566,
"grad_norm": 2.447306931764326,
"learning_rate": 1e-06,
"loss": 0.2585,
"step": 3723
},
{
"epoch": 1.4240917782026767,
"grad_norm": 3.166217136265214,
"learning_rate": 1e-06,
"loss": 0.1585,
"step": 3724
},
{
"epoch": 1.4244741873804971,
"grad_norm": 2.9064360528273805,
"learning_rate": 1e-06,
"loss": 0.1129,
"step": 3725
},
{
"epoch": 1.4248565965583175,
"grad_norm": 3.5005435521533594,
"learning_rate": 1e-06,
"loss": 0.1485,
"step": 3726
},
{
"epoch": 1.4252390057361377,
"grad_norm": 3.010948072651472,
"learning_rate": 1e-06,
"loss": 0.0942,
"step": 3727
},
{
"epoch": 1.4256214149139579,
"grad_norm": 6.661490369952854,
"learning_rate": 1e-06,
"loss": 0.1345,
"step": 3728
},
{
"epoch": 1.4260038240917783,
"grad_norm": 4.087868695444046,
"learning_rate": 1e-06,
"loss": 0.3328,
"step": 3729
},
{
"epoch": 1.4263862332695985,
"grad_norm": 4.022493062697949,
"learning_rate": 1e-06,
"loss": 0.4384,
"step": 3730
},
{
"epoch": 1.4267686424474189,
"grad_norm": 3.9391085048895844,
"learning_rate": 1e-06,
"loss": 0.268,
"step": 3731
},
{
"epoch": 1.427151051625239,
"grad_norm": 2.7178987069135747,
"learning_rate": 1e-06,
"loss": 0.2054,
"step": 3732
},
{
"epoch": 1.4275334608030592,
"grad_norm": 2.9878210174595545,
"learning_rate": 1e-06,
"loss": 0.2353,
"step": 3733
},
{
"epoch": 1.4279158699808796,
"grad_norm": 2.6006159162924245,
"learning_rate": 1e-06,
"loss": 0.2965,
"step": 3734
},
{
"epoch": 1.4282982791586998,
"grad_norm": 2.152111006266795,
"learning_rate": 1e-06,
"loss": 0.192,
"step": 3735
},
{
"epoch": 1.4286806883365202,
"grad_norm": 3.0866092771573768,
"learning_rate": 1e-06,
"loss": 0.1834,
"step": 3736
},
{
"epoch": 1.4290630975143404,
"grad_norm": 3.2257135119431815,
"learning_rate": 1e-06,
"loss": 0.0982,
"step": 3737
},
{
"epoch": 1.4294455066921605,
"grad_norm": 2.867569983641896,
"learning_rate": 1e-06,
"loss": 0.1885,
"step": 3738
},
{
"epoch": 1.429827915869981,
"grad_norm": 2.979210024075797,
"learning_rate": 1e-06,
"loss": 0.0917,
"step": 3739
},
{
"epoch": 1.430210325047801,
"grad_norm": 4.089296903731755,
"learning_rate": 1e-06,
"loss": 0.1143,
"step": 3740
},
{
"epoch": 1.4305927342256215,
"grad_norm": 6.8203038306252175,
"learning_rate": 1e-06,
"loss": 0.387,
"step": 3741
},
{
"epoch": 1.4309751434034417,
"grad_norm": 3.697388015712296,
"learning_rate": 1e-06,
"loss": 0.2882,
"step": 3742
},
{
"epoch": 1.4313575525812618,
"grad_norm": 3.100641147392317,
"learning_rate": 1e-06,
"loss": 0.2431,
"step": 3743
},
{
"epoch": 1.4317399617590822,
"grad_norm": 2.276666399824016,
"learning_rate": 1e-06,
"loss": 0.2894,
"step": 3744
},
{
"epoch": 1.4321223709369024,
"grad_norm": 3.471309370632797,
"learning_rate": 1e-06,
"loss": 0.4578,
"step": 3745
},
{
"epoch": 1.4325047801147228,
"grad_norm": 2.513064864700208,
"learning_rate": 1e-06,
"loss": 0.1828,
"step": 3746
},
{
"epoch": 1.432887189292543,
"grad_norm": 3.0271788324535414,
"learning_rate": 1e-06,
"loss": 0.1807,
"step": 3747
},
{
"epoch": 1.4332695984703632,
"grad_norm": 2.9643592890999306,
"learning_rate": 1e-06,
"loss": 0.1636,
"step": 3748
},
{
"epoch": 1.4336520076481836,
"grad_norm": 4.422537278901826,
"learning_rate": 1e-06,
"loss": 0.1512,
"step": 3749
},
{
"epoch": 1.4340344168260037,
"grad_norm": 5.567163805358466,
"learning_rate": 1e-06,
"loss": 0.1327,
"step": 3750
},
{
"epoch": 1.4344168260038241,
"grad_norm": 3.1560433197336115,
"learning_rate": 1e-06,
"loss": 0.1617,
"step": 3751
},
{
"epoch": 1.4347992351816443,
"grad_norm": 3.695903448552923,
"learning_rate": 1e-06,
"loss": 0.0543,
"step": 3752
},
{
"epoch": 1.4351816443594645,
"grad_norm": 3.4197311304602107,
"learning_rate": 1e-06,
"loss": 0.2029,
"step": 3753
},
{
"epoch": 1.435564053537285,
"grad_norm": 5.646728932571445,
"learning_rate": 1e-06,
"loss": 0.3716,
"step": 3754
},
{
"epoch": 1.4359464627151053,
"grad_norm": 3.627177143727946,
"learning_rate": 1e-06,
"loss": 0.4611,
"step": 3755
},
{
"epoch": 1.4363288718929255,
"grad_norm": 3.729141718321022,
"learning_rate": 1e-06,
"loss": 0.4277,
"step": 3756
},
{
"epoch": 1.4367112810707456,
"grad_norm": 1.9529400547201448,
"learning_rate": 1e-06,
"loss": 0.1611,
"step": 3757
},
{
"epoch": 1.437093690248566,
"grad_norm": 3.5934846199569557,
"learning_rate": 1e-06,
"loss": 0.3111,
"step": 3758
},
{
"epoch": 1.4374760994263862,
"grad_norm": 3.3197003686657816,
"learning_rate": 1e-06,
"loss": 0.2611,
"step": 3759
},
{
"epoch": 1.4378585086042066,
"grad_norm": 3.3328892094064173,
"learning_rate": 1e-06,
"loss": 0.1534,
"step": 3760
},
{
"epoch": 1.4382409177820268,
"grad_norm": 2.6049055551698954,
"learning_rate": 1e-06,
"loss": 0.1336,
"step": 3761
},
{
"epoch": 1.438623326959847,
"grad_norm": 2.225003308925954,
"learning_rate": 1e-06,
"loss": 0.0527,
"step": 3762
},
{
"epoch": 1.4390057361376674,
"grad_norm": 2.6469452291312163,
"learning_rate": 1e-06,
"loss": 0.0788,
"step": 3763
},
{
"epoch": 1.4393881453154875,
"grad_norm": 4.325611693063195,
"learning_rate": 1e-06,
"loss": 0.1896,
"step": 3764
},
{
"epoch": 1.439770554493308,
"grad_norm": 4.960876752826891,
"learning_rate": 1e-06,
"loss": 0.1686,
"step": 3765
},
{
"epoch": 1.440152963671128,
"grad_norm": 29.15869287394734,
"learning_rate": 1e-06,
"loss": 0.4078,
"step": 3766
},
{
"epoch": 1.4405353728489483,
"grad_norm": 4.755322335566404,
"learning_rate": 1e-06,
"loss": 0.4174,
"step": 3767
},
{
"epoch": 1.4409177820267687,
"grad_norm": 3.6924374386414796,
"learning_rate": 1e-06,
"loss": 0.1956,
"step": 3768
},
{
"epoch": 1.4413001912045889,
"grad_norm": 2.2364999763391546,
"learning_rate": 1e-06,
"loss": 0.2799,
"step": 3769
},
{
"epoch": 1.4416826003824093,
"grad_norm": 2.291046844816055,
"learning_rate": 1e-06,
"loss": 0.2377,
"step": 3770
},
{
"epoch": 1.4420650095602294,
"grad_norm": 2.9311425423192157,
"learning_rate": 1e-06,
"loss": 0.3293,
"step": 3771
},
{
"epoch": 1.4424474187380496,
"grad_norm": 4.293245315320849,
"learning_rate": 1e-06,
"loss": 0.1789,
"step": 3772
},
{
"epoch": 1.44282982791587,
"grad_norm": 3.8070113900377227,
"learning_rate": 1e-06,
"loss": 0.2407,
"step": 3773
},
{
"epoch": 1.4432122370936902,
"grad_norm": 3.4656250770506922,
"learning_rate": 1e-06,
"loss": 0.0927,
"step": 3774
},
{
"epoch": 1.4435946462715106,
"grad_norm": 2.83877917121469,
"learning_rate": 1e-06,
"loss": 0.1849,
"step": 3775
},
{
"epoch": 1.4439770554493307,
"grad_norm": 2.9861681279695182,
"learning_rate": 1e-06,
"loss": 0.1312,
"step": 3776
},
{
"epoch": 1.444359464627151,
"grad_norm": 4.455346981968557,
"learning_rate": 1e-06,
"loss": 0.1102,
"step": 3777
},
{
"epoch": 1.4447418738049713,
"grad_norm": 3.6213307719757655,
"learning_rate": 1e-06,
"loss": 0.0999,
"step": 3778
},
{
"epoch": 1.4451242829827917,
"grad_norm": 3.864392682879823,
"learning_rate": 1e-06,
"loss": 0.4747,
"step": 3779
},
{
"epoch": 1.445506692160612,
"grad_norm": 3.804148171676907,
"learning_rate": 1e-06,
"loss": 0.3398,
"step": 3780
},
{
"epoch": 1.445889101338432,
"grad_norm": 2.6989292812121124,
"learning_rate": 1e-06,
"loss": 0.1673,
"step": 3781
},
{
"epoch": 1.4462715105162525,
"grad_norm": 2.8561388056707435,
"learning_rate": 1e-06,
"loss": 0.2778,
"step": 3782
},
{
"epoch": 1.4466539196940726,
"grad_norm": 3.971278549058097,
"learning_rate": 1e-06,
"loss": 0.402,
"step": 3783
},
{
"epoch": 1.447036328871893,
"grad_norm": 2.9940332840144133,
"learning_rate": 1e-06,
"loss": 0.274,
"step": 3784
},
{
"epoch": 1.4474187380497132,
"grad_norm": 3.310150464835929,
"learning_rate": 1e-06,
"loss": 0.1365,
"step": 3785
},
{
"epoch": 1.4478011472275334,
"grad_norm": 2.8408065486088017,
"learning_rate": 1e-06,
"loss": 0.0925,
"step": 3786
},
{
"epoch": 1.4481835564053538,
"grad_norm": 3.0926545640933805,
"learning_rate": 1e-06,
"loss": 0.1657,
"step": 3787
},
{
"epoch": 1.448565965583174,
"grad_norm": 2.6221871518719135,
"learning_rate": 1e-06,
"loss": 0.0585,
"step": 3788
},
{
"epoch": 1.4489483747609944,
"grad_norm": 2.4402629158765445,
"learning_rate": 1e-06,
"loss": 0.0521,
"step": 3789
},
{
"epoch": 1.4493307839388145,
"grad_norm": 3.840107378054394,
"learning_rate": 1e-06,
"loss": 0.1009,
"step": 3790
},
{
"epoch": 1.4497131931166347,
"grad_norm": 7.8042168002768,
"learning_rate": 1e-06,
"loss": 0.4154,
"step": 3791
},
{
"epoch": 1.450095602294455,
"grad_norm": 3.7882124429040727,
"learning_rate": 1e-06,
"loss": 0.3955,
"step": 3792
},
{
"epoch": 1.4504780114722753,
"grad_norm": 4.218763111235413,
"learning_rate": 1e-06,
"loss": 0.5727,
"step": 3793
},
{
"epoch": 1.4508604206500957,
"grad_norm": 2.3773355544667005,
"learning_rate": 1e-06,
"loss": 0.1635,
"step": 3794
},
{
"epoch": 1.4512428298279159,
"grad_norm": 2.944363145085735,
"learning_rate": 1e-06,
"loss": 0.2911,
"step": 3795
},
{
"epoch": 1.451625239005736,
"grad_norm": 3.8709171073944915,
"learning_rate": 1e-06,
"loss": 0.4245,
"step": 3796
},
{
"epoch": 1.4520076481835564,
"grad_norm": 3.4701117514608883,
"learning_rate": 1e-06,
"loss": 0.3219,
"step": 3797
},
{
"epoch": 1.4523900573613766,
"grad_norm": 4.621335820543163,
"learning_rate": 1e-06,
"loss": 0.2913,
"step": 3798
},
{
"epoch": 1.452772466539197,
"grad_norm": 2.8795663809181855,
"learning_rate": 1e-06,
"loss": 0.1822,
"step": 3799
},
{
"epoch": 1.4531548757170172,
"grad_norm": 2.873432561425201,
"learning_rate": 1e-06,
"loss": 0.1353,
"step": 3800
},
{
"epoch": 1.4535372848948374,
"grad_norm": 3.335319515404248,
"learning_rate": 1e-06,
"loss": 0.1563,
"step": 3801
},
{
"epoch": 1.4539196940726578,
"grad_norm": 4.237311101883541,
"learning_rate": 1e-06,
"loss": 0.0742,
"step": 3802
},
{
"epoch": 1.454302103250478,
"grad_norm": 6.744611496648465,
"learning_rate": 1e-06,
"loss": 0.2078,
"step": 3803
},
{
"epoch": 1.4546845124282983,
"grad_norm": 3.9523890115499265,
"learning_rate": 1e-06,
"loss": 0.391,
"step": 3804
},
{
"epoch": 1.4550669216061185,
"grad_norm": 4.015880769532698,
"learning_rate": 1e-06,
"loss": 0.5503,
"step": 3805
},
{
"epoch": 1.4554493307839387,
"grad_norm": 3.066642450845172,
"learning_rate": 1e-06,
"loss": 0.2996,
"step": 3806
},
{
"epoch": 1.455831739961759,
"grad_norm": 3.2684017951430984,
"learning_rate": 1e-06,
"loss": 0.4948,
"step": 3807
},
{
"epoch": 1.4562141491395795,
"grad_norm": 3.300095418070754,
"learning_rate": 1e-06,
"loss": 0.2781,
"step": 3808
},
{
"epoch": 1.4565965583173996,
"grad_norm": 3.419785092165752,
"learning_rate": 1e-06,
"loss": 0.1961,
"step": 3809
},
{
"epoch": 1.4569789674952198,
"grad_norm": 2.9591561547703815,
"learning_rate": 1e-06,
"loss": 0.3269,
"step": 3810
},
{
"epoch": 1.4573613766730402,
"grad_norm": 3.1471698944808435,
"learning_rate": 1e-06,
"loss": 0.1619,
"step": 3811
},
{
"epoch": 1.4577437858508604,
"grad_norm": 3.00136233232753,
"learning_rate": 1e-06,
"loss": 0.1578,
"step": 3812
},
{
"epoch": 1.4581261950286808,
"grad_norm": 3.375788561215613,
"learning_rate": 1e-06,
"loss": 0.2613,
"step": 3813
},
{
"epoch": 1.458508604206501,
"grad_norm": 2.9304037210464284,
"learning_rate": 1e-06,
"loss": 0.086,
"step": 3814
},
{
"epoch": 1.4588910133843211,
"grad_norm": 3.3510198264951336,
"learning_rate": 1e-06,
"loss": 0.1134,
"step": 3815
},
{
"epoch": 1.4592734225621415,
"grad_norm": 3.97487978003701,
"learning_rate": 1e-06,
"loss": 0.4668,
"step": 3816
},
{
"epoch": 1.4596558317399617,
"grad_norm": 3.71832054928221,
"learning_rate": 1e-06,
"loss": 0.4395,
"step": 3817
},
{
"epoch": 1.4600382409177821,
"grad_norm": 4.98342743001193,
"learning_rate": 1e-06,
"loss": 0.5187,
"step": 3818
},
{
"epoch": 1.4604206500956023,
"grad_norm": 3.080457239518899,
"learning_rate": 1e-06,
"loss": 0.2241,
"step": 3819
},
{
"epoch": 1.4608030592734225,
"grad_norm": 4.7729919525340785,
"learning_rate": 1e-06,
"loss": 0.2974,
"step": 3820
},
{
"epoch": 1.4611854684512429,
"grad_norm": 3.419843375428673,
"learning_rate": 1e-06,
"loss": 0.2616,
"step": 3821
},
{
"epoch": 1.461567877629063,
"grad_norm": 2.9801685987052786,
"learning_rate": 1e-06,
"loss": 0.2581,
"step": 3822
},
{
"epoch": 1.4619502868068834,
"grad_norm": 2.9065922361287666,
"learning_rate": 1e-06,
"loss": 0.168,
"step": 3823
},
{
"epoch": 1.4623326959847036,
"grad_norm": 3.7695722350427348,
"learning_rate": 1e-06,
"loss": 0.152,
"step": 3824
},
{
"epoch": 1.4627151051625238,
"grad_norm": 3.174334394717624,
"learning_rate": 1e-06,
"loss": 0.1804,
"step": 3825
},
{
"epoch": 1.4630975143403442,
"grad_norm": 4.837848045583718,
"learning_rate": 1e-06,
"loss": 0.1235,
"step": 3826
},
{
"epoch": 1.4634799235181644,
"grad_norm": 3.2905538263258594,
"learning_rate": 1e-06,
"loss": 0.0845,
"step": 3827
},
{
"epoch": 1.4638623326959848,
"grad_norm": 9.019931762843202,
"learning_rate": 1e-06,
"loss": 0.1512,
"step": 3828
},
{
"epoch": 1.464244741873805,
"grad_norm": 4.02266494592974,
"learning_rate": 1e-06,
"loss": 0.4447,
"step": 3829
},
{
"epoch": 1.464627151051625,
"grad_norm": 3.7410718811846464,
"learning_rate": 1e-06,
"loss": 0.2839,
"step": 3830
},
{
"epoch": 1.4650095602294455,
"grad_norm": 2.7484205218338853,
"learning_rate": 1e-06,
"loss": 0.1299,
"step": 3831
},
{
"epoch": 1.465391969407266,
"grad_norm": 3.654099573775783,
"learning_rate": 1e-06,
"loss": 0.3325,
"step": 3832
},
{
"epoch": 1.465774378585086,
"grad_norm": 2.777116940622431,
"learning_rate": 1e-06,
"loss": 0.2626,
"step": 3833
},
{
"epoch": 1.4661567877629063,
"grad_norm": 3.1262529531630423,
"learning_rate": 1e-06,
"loss": 0.1753,
"step": 3834
},
{
"epoch": 1.4665391969407267,
"grad_norm": 2.904293427147836,
"learning_rate": 1e-06,
"loss": 0.1693,
"step": 3835
},
{
"epoch": 1.4669216061185468,
"grad_norm": 2.8254094865133252,
"learning_rate": 1e-06,
"loss": 0.175,
"step": 3836
},
{
"epoch": 1.4673040152963672,
"grad_norm": 4.781284307219646,
"learning_rate": 1e-06,
"loss": 0.1232,
"step": 3837
},
{
"epoch": 1.4676864244741874,
"grad_norm": 3.1129096562075103,
"learning_rate": 1e-06,
"loss": 0.0863,
"step": 3838
},
{
"epoch": 1.4680688336520076,
"grad_norm": 4.842424568215032,
"learning_rate": 1e-06,
"loss": 0.1054,
"step": 3839
},
{
"epoch": 1.468451242829828,
"grad_norm": 3.693478667593567,
"learning_rate": 1e-06,
"loss": 0.105,
"step": 3840
},
{
"epoch": 1.4688336520076481,
"grad_norm": 8.429961729494517,
"learning_rate": 1e-06,
"loss": 0.3211,
"step": 3841
},
{
"epoch": 1.4692160611854685,
"grad_norm": 3.267691364594895,
"learning_rate": 1e-06,
"loss": 0.2938,
"step": 3842
},
{
"epoch": 1.4695984703632887,
"grad_norm": 3.8187200117416533,
"learning_rate": 1e-06,
"loss": 0.3805,
"step": 3843
},
{
"epoch": 1.469980879541109,
"grad_norm": 3.5728148658122003,
"learning_rate": 1e-06,
"loss": 0.2267,
"step": 3844
},
{
"epoch": 1.4703632887189293,
"grad_norm": 3.0617581461566554,
"learning_rate": 1e-06,
"loss": 0.2212,
"step": 3845
},
{
"epoch": 1.4707456978967495,
"grad_norm": 2.946903517454639,
"learning_rate": 1e-06,
"loss": 0.3188,
"step": 3846
},
{
"epoch": 1.4711281070745699,
"grad_norm": 2.6753001427941547,
"learning_rate": 1e-06,
"loss": 0.1129,
"step": 3847
},
{
"epoch": 1.47151051625239,
"grad_norm": 2.5027372633796934,
"learning_rate": 1e-06,
"loss": 0.0802,
"step": 3848
},
{
"epoch": 1.4718929254302102,
"grad_norm": 4.395501195041953,
"learning_rate": 1e-06,
"loss": 0.15,
"step": 3849
},
{
"epoch": 1.4722753346080306,
"grad_norm": 3.773209527436749,
"learning_rate": 1e-06,
"loss": 0.1679,
"step": 3850
},
{
"epoch": 1.4726577437858508,
"grad_norm": 3.860204445361068,
"learning_rate": 1e-06,
"loss": 0.1024,
"step": 3851
},
{
"epoch": 1.4730401529636712,
"grad_norm": 3.530051509059919,
"learning_rate": 1e-06,
"loss": 0.0615,
"step": 3852
},
{
"epoch": 1.4734225621414914,
"grad_norm": 5.866956033217151,
"learning_rate": 1e-06,
"loss": 0.1333,
"step": 3853
},
{
"epoch": 1.4738049713193115,
"grad_norm": 6.358621238633143,
"learning_rate": 1e-06,
"loss": 0.5757,
"step": 3854
},
{
"epoch": 1.474187380497132,
"grad_norm": 4.513029943399394,
"learning_rate": 1e-06,
"loss": 0.4743,
"step": 3855
},
{
"epoch": 1.4745697896749523,
"grad_norm": 2.8871868409602897,
"learning_rate": 1e-06,
"loss": 0.247,
"step": 3856
},
{
"epoch": 1.4749521988527725,
"grad_norm": 3.2121236795957935,
"learning_rate": 1e-06,
"loss": 0.1984,
"step": 3857
},
{
"epoch": 1.4753346080305927,
"grad_norm": 2.2729066838628866,
"learning_rate": 1e-06,
"loss": 0.2743,
"step": 3858
},
{
"epoch": 1.475717017208413,
"grad_norm": 4.660401407397884,
"learning_rate": 1e-06,
"loss": 0.334,
"step": 3859
},
{
"epoch": 1.4760994263862333,
"grad_norm": 3.0518740602859404,
"learning_rate": 1e-06,
"loss": 0.2125,
"step": 3860
},
{
"epoch": 1.4764818355640537,
"grad_norm": 2.2497827636964605,
"learning_rate": 1e-06,
"loss": 0.1154,
"step": 3861
},
{
"epoch": 1.4768642447418738,
"grad_norm": 2.755924951326017,
"learning_rate": 1e-06,
"loss": 0.0984,
"step": 3862
},
{
"epoch": 1.477246653919694,
"grad_norm": 3.4864116925821222,
"learning_rate": 1e-06,
"loss": 0.1778,
"step": 3863
},
{
"epoch": 1.4776290630975144,
"grad_norm": 3.5730174562939037,
"learning_rate": 1e-06,
"loss": 0.0803,
"step": 3864
},
{
"epoch": 1.4780114722753346,
"grad_norm": 3.129699225553734,
"learning_rate": 1e-06,
"loss": 0.0986,
"step": 3865
},
{
"epoch": 1.478393881453155,
"grad_norm": 3.894855328910445,
"learning_rate": 1e-06,
"loss": 0.3113,
"step": 3866
},
{
"epoch": 1.4787762906309752,
"grad_norm": 3.941234452555413,
"learning_rate": 1e-06,
"loss": 0.3866,
"step": 3867
},
{
"epoch": 1.4791586998087953,
"grad_norm": 5.609048222876385,
"learning_rate": 1e-06,
"loss": 0.3934,
"step": 3868
},
{
"epoch": 1.4795411089866157,
"grad_norm": 2.3695421248179196,
"learning_rate": 1e-06,
"loss": 0.2293,
"step": 3869
},
{
"epoch": 1.479923518164436,
"grad_norm": 3.562778261259576,
"learning_rate": 1e-06,
"loss": 0.259,
"step": 3870
},
{
"epoch": 1.4803059273422563,
"grad_norm": 3.198581506331975,
"learning_rate": 1e-06,
"loss": 0.229,
"step": 3871
},
{
"epoch": 1.4806883365200765,
"grad_norm": 2.1479781335964403,
"learning_rate": 1e-06,
"loss": 0.1967,
"step": 3872
},
{
"epoch": 1.4810707456978967,
"grad_norm": 3.325525549176048,
"learning_rate": 1e-06,
"loss": 0.1255,
"step": 3873
},
{
"epoch": 1.481453154875717,
"grad_norm": 4.47129439897385,
"learning_rate": 1e-06,
"loss": 0.1792,
"step": 3874
},
{
"epoch": 1.4818355640535372,
"grad_norm": 3.019859069163232,
"learning_rate": 1e-06,
"loss": 0.1168,
"step": 3875
},
{
"epoch": 1.4822179732313576,
"grad_norm": 4.957213245895642,
"learning_rate": 1e-06,
"loss": 0.1112,
"step": 3876
},
{
"epoch": 1.4826003824091778,
"grad_norm": 3.5225665248764537,
"learning_rate": 1e-06,
"loss": 0.0608,
"step": 3877
},
{
"epoch": 1.482982791586998,
"grad_norm": 5.737249417951869,
"learning_rate": 1e-06,
"loss": 0.1176,
"step": 3878
},
{
"epoch": 1.4833652007648184,
"grad_norm": 3.5995974527576573,
"learning_rate": 1e-06,
"loss": 0.3218,
"step": 3879
},
{
"epoch": 1.4837476099426385,
"grad_norm": 4.423747714156061,
"learning_rate": 1e-06,
"loss": 0.4048,
"step": 3880
},
{
"epoch": 1.484130019120459,
"grad_norm": 2.0583476257857614,
"learning_rate": 1e-06,
"loss": 0.2074,
"step": 3881
},
{
"epoch": 1.4845124282982791,
"grad_norm": 3.4366762041119294,
"learning_rate": 1e-06,
"loss": 0.1247,
"step": 3882
},
{
"epoch": 1.4848948374760993,
"grad_norm": 3.6979692195686207,
"learning_rate": 1e-06,
"loss": 0.298,
"step": 3883
},
{
"epoch": 1.4852772466539197,
"grad_norm": 2.3355698312108766,
"learning_rate": 1e-06,
"loss": 0.1487,
"step": 3884
},
{
"epoch": 1.48565965583174,
"grad_norm": 1.807286492449328,
"learning_rate": 1e-06,
"loss": 0.0802,
"step": 3885
},
{
"epoch": 1.4860420650095603,
"grad_norm": 3.6630470153807266,
"learning_rate": 1e-06,
"loss": 0.2576,
"step": 3886
},
{
"epoch": 1.4864244741873804,
"grad_norm": 2.801169566710892,
"learning_rate": 1e-06,
"loss": 0.0663,
"step": 3887
},
{
"epoch": 1.4868068833652008,
"grad_norm": 2.9192060224613656,
"learning_rate": 1e-06,
"loss": 0.1058,
"step": 3888
},
{
"epoch": 1.487189292543021,
"grad_norm": 2.5053213705995705,
"learning_rate": 1e-06,
"loss": 0.069,
"step": 3889
},
{
"epoch": 1.4875717017208414,
"grad_norm": 4.644347268223121,
"learning_rate": 1e-06,
"loss": 0.0938,
"step": 3890
},
{
"epoch": 1.4879541108986616,
"grad_norm": 5.329618272692444,
"learning_rate": 1e-06,
"loss": 0.4061,
"step": 3891
},
{
"epoch": 1.4883365200764818,
"grad_norm": 4.072459527757717,
"learning_rate": 1e-06,
"loss": 0.4351,
"step": 3892
},
{
"epoch": 1.4887189292543022,
"grad_norm": 3.5818813546370136,
"learning_rate": 1e-06,
"loss": 0.2766,
"step": 3893
},
{
"epoch": 1.4891013384321223,
"grad_norm": 2.5468588986502154,
"learning_rate": 1e-06,
"loss": 0.283,
"step": 3894
},
{
"epoch": 1.4894837476099427,
"grad_norm": 2.7777776580386666,
"learning_rate": 1e-06,
"loss": 0.4183,
"step": 3895
},
{
"epoch": 1.489866156787763,
"grad_norm": 3.3154771489607286,
"learning_rate": 1e-06,
"loss": 0.2987,
"step": 3896
},
{
"epoch": 1.490248565965583,
"grad_norm": 3.2799520195963985,
"learning_rate": 1e-06,
"loss": 0.1745,
"step": 3897
},
{
"epoch": 1.4906309751434035,
"grad_norm": 2.9072710110076274,
"learning_rate": 1e-06,
"loss": 0.1737,
"step": 3898
},
{
"epoch": 1.4910133843212237,
"grad_norm": 2.999697034160772,
"learning_rate": 1e-06,
"loss": 0.2543,
"step": 3899
},
{
"epoch": 1.491395793499044,
"grad_norm": 2.356610124423201,
"learning_rate": 1e-06,
"loss": 0.1466,
"step": 3900
},
{
"epoch": 1.4917782026768642,
"grad_norm": 3.3811958802240865,
"learning_rate": 1e-06,
"loss": 0.1331,
"step": 3901
},
{
"epoch": 1.4921606118546844,
"grad_norm": 3.9710000936123433,
"learning_rate": 1e-06,
"loss": 0.0954,
"step": 3902
},
{
"epoch": 1.4925430210325048,
"grad_norm": 6.990058105838466,
"learning_rate": 1e-06,
"loss": 0.2311,
"step": 3903
},
{
"epoch": 1.492925430210325,
"grad_norm": 4.267945716791568,
"learning_rate": 1e-06,
"loss": 0.4079,
"step": 3904
},
{
"epoch": 1.4933078393881454,
"grad_norm": 3.1010631091976446,
"learning_rate": 1e-06,
"loss": 0.3872,
"step": 3905
},
{
"epoch": 1.4936902485659656,
"grad_norm": 3.8123869331978844,
"learning_rate": 1e-06,
"loss": 0.2134,
"step": 3906
},
{
"epoch": 1.4940726577437857,
"grad_norm": 3.3967965910985645,
"learning_rate": 1e-06,
"loss": 0.1731,
"step": 3907
},
{
"epoch": 1.4944550669216061,
"grad_norm": 3.937442476367078,
"learning_rate": 1e-06,
"loss": 0.3576,
"step": 3908
},
{
"epoch": 1.4948374760994265,
"grad_norm": 2.028552332736425,
"learning_rate": 1e-06,
"loss": 0.1715,
"step": 3909
},
{
"epoch": 1.4952198852772467,
"grad_norm": 2.9499790449529675,
"learning_rate": 1e-06,
"loss": 0.1708,
"step": 3910
},
{
"epoch": 1.4956022944550669,
"grad_norm": 4.119911754737143,
"learning_rate": 1e-06,
"loss": 0.1522,
"step": 3911
},
{
"epoch": 1.4959847036328873,
"grad_norm": 3.274174500304965,
"learning_rate": 1e-06,
"loss": 0.1774,
"step": 3912
},
{
"epoch": 1.4963671128107074,
"grad_norm": 3.4899167181533137,
"learning_rate": 1e-06,
"loss": 0.1361,
"step": 3913
},
{
"epoch": 1.4967495219885278,
"grad_norm": 4.463253882918679,
"learning_rate": 1e-06,
"loss": 0.1868,
"step": 3914
},
{
"epoch": 1.497131931166348,
"grad_norm": 3.775710911789235,
"learning_rate": 1e-06,
"loss": 0.0805,
"step": 3915
},
{
"epoch": 1.4975143403441682,
"grad_norm": 6.183630455809408,
"learning_rate": 1e-06,
"loss": 0.2829,
"step": 3916
},
{
"epoch": 1.4978967495219886,
"grad_norm": 3.0684675346945762,
"learning_rate": 1e-06,
"loss": 0.4096,
"step": 3917
},
{
"epoch": 1.4982791586998088,
"grad_norm": 3.9605638075693372,
"learning_rate": 1e-06,
"loss": 0.3544,
"step": 3918
},
{
"epoch": 1.4986615678776292,
"grad_norm": 2.305883168558684,
"learning_rate": 1e-06,
"loss": 0.2506,
"step": 3919
},
{
"epoch": 1.4990439770554493,
"grad_norm": 2.8661289183856753,
"learning_rate": 1e-06,
"loss": 0.2661,
"step": 3920
},
{
"epoch": 1.4994263862332695,
"grad_norm": 3.0950222289438005,
"learning_rate": 1e-06,
"loss": 0.3595,
"step": 3921
},
{
"epoch": 1.49980879541109,
"grad_norm": 3.3450728322127183,
"learning_rate": 1e-06,
"loss": 0.194,
"step": 3922
},
{
"epoch": 1.50019120458891,
"grad_norm": 2.816971932371236,
"learning_rate": 1e-06,
"loss": 0.151,
"step": 3923
},
{
"epoch": 1.5005736137667305,
"grad_norm": 5.180212210060488,
"learning_rate": 1e-06,
"loss": 0.2595,
"step": 3924
},
{
"epoch": 1.5009560229445507,
"grad_norm": 4.4262759680472366,
"learning_rate": 1e-06,
"loss": 0.1916,
"step": 3925
},
{
"epoch": 1.5013384321223708,
"grad_norm": 5.305925497280416,
"learning_rate": 1e-06,
"loss": 0.1748,
"step": 3926
},
{
"epoch": 1.5017208413001912,
"grad_norm": 6.767331231968318,
"learning_rate": 1e-06,
"loss": 0.0759,
"step": 3927
},
{
"epoch": 1.5021032504780116,
"grad_norm": 4.925718907590221,
"learning_rate": 1e-06,
"loss": 0.1051,
"step": 3928
},
{
"epoch": 1.5024856596558318,
"grad_norm": 4.954566914159578,
"learning_rate": 1e-06,
"loss": 0.4665,
"step": 3929
},
{
"epoch": 1.502868068833652,
"grad_norm": 2.9394148409051746,
"learning_rate": 1e-06,
"loss": 0.3569,
"step": 3930
},
{
"epoch": 1.5032504780114722,
"grad_norm": 2.4671448909856215,
"learning_rate": 1e-06,
"loss": 0.1468,
"step": 3931
},
{
"epoch": 1.5036328871892926,
"grad_norm": 2.6880407454717656,
"learning_rate": 1e-06,
"loss": 0.2397,
"step": 3932
},
{
"epoch": 1.504015296367113,
"grad_norm": 4.153998891081867,
"learning_rate": 1e-06,
"loss": 0.2802,
"step": 3933
},
{
"epoch": 1.5043977055449331,
"grad_norm": 2.1257717750615743,
"learning_rate": 1e-06,
"loss": 0.1334,
"step": 3934
},
{
"epoch": 1.5047801147227533,
"grad_norm": 3.097527403875748,
"learning_rate": 1e-06,
"loss": 0.2852,
"step": 3935
},
{
"epoch": 1.5051625239005735,
"grad_norm": 2.4833219203644044,
"learning_rate": 1e-06,
"loss": 0.1797,
"step": 3936
},
{
"epoch": 1.5055449330783939,
"grad_norm": 1.8667316598590287,
"learning_rate": 1e-06,
"loss": 0.0644,
"step": 3937
},
{
"epoch": 1.5059273422562143,
"grad_norm": 3.546294177388295,
"learning_rate": 1e-06,
"loss": 0.161,
"step": 3938
},
{
"epoch": 1.5063097514340344,
"grad_norm": 2.9370517794571485,
"learning_rate": 1e-06,
"loss": 0.122,
"step": 3939
},
{
"epoch": 1.5066921606118546,
"grad_norm": 4.558320521883538,
"learning_rate": 1e-06,
"loss": 0.1002,
"step": 3940
},
{
"epoch": 1.5070745697896748,
"grad_norm": 4.809699891179583,
"learning_rate": 1e-06,
"loss": 0.54,
"step": 3941
},
{
"epoch": 1.5074569789674952,
"grad_norm": 3.744581376680773,
"learning_rate": 1e-06,
"loss": 0.4046,
"step": 3942
},
{
"epoch": 1.5078393881453156,
"grad_norm": 7.3905973857339555,
"learning_rate": 1e-06,
"loss": 0.4947,
"step": 3943
},
{
"epoch": 1.5082217973231358,
"grad_norm": 2.8259975798772667,
"learning_rate": 1e-06,
"loss": 0.2338,
"step": 3944
},
{
"epoch": 1.508604206500956,
"grad_norm": 2.8417804311165136,
"learning_rate": 1e-06,
"loss": 0.1986,
"step": 3945
},
{
"epoch": 1.5089866156787763,
"grad_norm": 2.783702840281398,
"learning_rate": 1e-06,
"loss": 0.3213,
"step": 3946
},
{
"epoch": 1.5093690248565965,
"grad_norm": 2.23992316454852,
"learning_rate": 1e-06,
"loss": 0.128,
"step": 3947
},
{
"epoch": 1.509751434034417,
"grad_norm": 2.86368157267148,
"learning_rate": 1e-06,
"loss": 0.2186,
"step": 3948
},
{
"epoch": 1.510133843212237,
"grad_norm": 2.986499289977013,
"learning_rate": 1e-06,
"loss": 0.1047,
"step": 3949
},
{
"epoch": 1.5105162523900573,
"grad_norm": 2.602839709123787,
"learning_rate": 1e-06,
"loss": 0.0556,
"step": 3950
},
{
"epoch": 1.5108986615678777,
"grad_norm": 3.4712862931999586,
"learning_rate": 1e-06,
"loss": 0.1932,
"step": 3951
},
{
"epoch": 1.511281070745698,
"grad_norm": 4.825550170523791,
"learning_rate": 1e-06,
"loss": 0.0981,
"step": 3952
},
{
"epoch": 1.5116634799235182,
"grad_norm": 5.906107582918369,
"learning_rate": 1e-06,
"loss": 0.1945,
"step": 3953
},
{
"epoch": 1.5120458891013384,
"grad_norm": 4.152000147472225,
"learning_rate": 1e-06,
"loss": 0.6537,
"step": 3954
},
{
"epoch": 1.5124282982791586,
"grad_norm": 2.9336616924763703,
"learning_rate": 1e-06,
"loss": 0.3252,
"step": 3955
},
{
"epoch": 1.512810707456979,
"grad_norm": 1.758289865736911,
"learning_rate": 1e-06,
"loss": 0.2387,
"step": 3956
},
{
"epoch": 1.5131931166347994,
"grad_norm": 4.006370478349455,
"learning_rate": 1e-06,
"loss": 0.2336,
"step": 3957
},
{
"epoch": 1.5135755258126196,
"grad_norm": 3.675672020743574,
"learning_rate": 1e-06,
"loss": 0.2642,
"step": 3958
},
{
"epoch": 1.5139579349904397,
"grad_norm": 3.4303843978487882,
"learning_rate": 1e-06,
"loss": 0.1566,
"step": 3959
},
{
"epoch": 1.51434034416826,
"grad_norm": 2.675526583170763,
"learning_rate": 1e-06,
"loss": 0.115,
"step": 3960
},
{
"epoch": 1.5147227533460803,
"grad_norm": 3.04067754121807,
"learning_rate": 1e-06,
"loss": 0.1148,
"step": 3961
},
{
"epoch": 1.5151051625239007,
"grad_norm": 4.519021416792543,
"learning_rate": 1e-06,
"loss": 0.1472,
"step": 3962
},
{
"epoch": 1.5154875717017209,
"grad_norm": 3.2548776850300896,
"learning_rate": 1e-06,
"loss": 0.1106,
"step": 3963
},
{
"epoch": 1.515869980879541,
"grad_norm": 3.9135161911344487,
"learning_rate": 1e-06,
"loss": 0.087,
"step": 3964
},
{
"epoch": 1.5162523900573612,
"grad_norm": 4.010244839773097,
"learning_rate": 1e-06,
"loss": 0.1205,
"step": 3965
},
{
"epoch": 1.5166347992351816,
"grad_norm": 4.56472932880087,
"learning_rate": 1e-06,
"loss": 0.4076,
"step": 3966
},
{
"epoch": 1.517017208413002,
"grad_norm": 5.479372008601221,
"learning_rate": 1e-06,
"loss": 0.4973,
"step": 3967
},
{
"epoch": 1.5173996175908222,
"grad_norm": 4.299789494529427,
"learning_rate": 1e-06,
"loss": 0.4959,
"step": 3968
},
{
"epoch": 1.5177820267686424,
"grad_norm": 2.925059821666413,
"learning_rate": 1e-06,
"loss": 0.1516,
"step": 3969
},
{
"epoch": 1.5181644359464626,
"grad_norm": 2.8787240497381847,
"learning_rate": 1e-06,
"loss": 0.2397,
"step": 3970
},
{
"epoch": 1.518546845124283,
"grad_norm": 3.7031709692310937,
"learning_rate": 1e-06,
"loss": 0.2515,
"step": 3971
},
{
"epoch": 1.5189292543021033,
"grad_norm": 1.955050686425526,
"learning_rate": 1e-06,
"loss": 0.1695,
"step": 3972
},
{
"epoch": 1.5193116634799235,
"grad_norm": 3.113865662290679,
"learning_rate": 1e-06,
"loss": 0.1802,
"step": 3973
},
{
"epoch": 1.5196940726577437,
"grad_norm": 3.2465922756667407,
"learning_rate": 1e-06,
"loss": 0.2266,
"step": 3974
},
{
"epoch": 1.520076481835564,
"grad_norm": 2.9431008046262033,
"learning_rate": 1e-06,
"loss": 0.1225,
"step": 3975
},
{
"epoch": 1.5204588910133843,
"grad_norm": 3.030820993169864,
"learning_rate": 1e-06,
"loss": 0.0523,
"step": 3976
},
{
"epoch": 1.5208413001912047,
"grad_norm": 3.3309977137094817,
"learning_rate": 1e-06,
"loss": 0.0825,
"step": 3977
},
{
"epoch": 1.5212237093690248,
"grad_norm": 35.00118231819998,
"learning_rate": 1e-06,
"loss": 0.2887,
"step": 3978
},
{
"epoch": 1.521606118546845,
"grad_norm": 5.631251950956932,
"learning_rate": 1e-06,
"loss": 0.3441,
"step": 3979
},
{
"epoch": 1.5219885277246654,
"grad_norm": 3.851298046511904,
"learning_rate": 1e-06,
"loss": 0.3802,
"step": 3980
},
{
"epoch": 1.5223709369024858,
"grad_norm": 4.1656829944416875,
"learning_rate": 1e-06,
"loss": 0.4271,
"step": 3981
},
{
"epoch": 1.522753346080306,
"grad_norm": 2.903704441089205,
"learning_rate": 1e-06,
"loss": 0.2593,
"step": 3982
},
{
"epoch": 1.5231357552581262,
"grad_norm": 2.0742226580854193,
"learning_rate": 1e-06,
"loss": 0.1707,
"step": 3983
},
{
"epoch": 1.5235181644359463,
"grad_norm": 3.663700175315848,
"learning_rate": 1e-06,
"loss": 0.2572,
"step": 3984
},
{
"epoch": 1.5239005736137667,
"grad_norm": 2.596237460580444,
"learning_rate": 1e-06,
"loss": 0.0764,
"step": 3985
},
{
"epoch": 1.5242829827915871,
"grad_norm": 3.001227445632857,
"learning_rate": 1e-06,
"loss": 0.1724,
"step": 3986
},
{
"epoch": 1.5246653919694073,
"grad_norm": 3.2115662045117355,
"learning_rate": 1e-06,
"loss": 0.1363,
"step": 3987
},
{
"epoch": 1.5250478011472275,
"grad_norm": 2.402127189489529,
"learning_rate": 1e-06,
"loss": 0.0842,
"step": 3988
},
{
"epoch": 1.5254302103250477,
"grad_norm": 3.184964836082194,
"learning_rate": 1e-06,
"loss": 0.0764,
"step": 3989
},
{
"epoch": 1.525812619502868,
"grad_norm": 4.7110279890640765,
"learning_rate": 1e-06,
"loss": 0.2332,
"step": 3990
},
{
"epoch": 1.5261950286806885,
"grad_norm": 6.3784233326273565,
"learning_rate": 1e-06,
"loss": 0.1823,
"step": 3991
},
{
"epoch": 1.5265774378585086,
"grad_norm": 5.270771533870147,
"learning_rate": 1e-06,
"loss": 0.5324,
"step": 3992
},
{
"epoch": 1.5269598470363288,
"grad_norm": 2.952825298945698,
"learning_rate": 1e-06,
"loss": 0.339,
"step": 3993
},
{
"epoch": 1.527342256214149,
"grad_norm": 4.203740298248357,
"learning_rate": 1e-06,
"loss": 0.2524,
"step": 3994
},
{
"epoch": 1.5277246653919694,
"grad_norm": 3.2305146551964388,
"learning_rate": 1e-06,
"loss": 0.2367,
"step": 3995
},
{
"epoch": 1.5281070745697898,
"grad_norm": 3.0066196206493974,
"learning_rate": 1e-06,
"loss": 0.4437,
"step": 3996
},
{
"epoch": 1.52848948374761,
"grad_norm": 3.164315823433579,
"learning_rate": 1e-06,
"loss": 0.1544,
"step": 3997
},
{
"epoch": 1.5288718929254301,
"grad_norm": 2.617455676558822,
"learning_rate": 1e-06,
"loss": 0.1725,
"step": 3998
},
{
"epoch": 1.5292543021032505,
"grad_norm": 3.616000085628137,
"learning_rate": 1e-06,
"loss": 0.2934,
"step": 3999
},
{
"epoch": 1.5296367112810707,
"grad_norm": 3.760712137352352,
"learning_rate": 1e-06,
"loss": 0.1458,
"step": 4000
},
{
"epoch": 1.5296367112810707,
"eval_runtime": 840.9805,
"eval_samples_per_second": 1.824,
"eval_steps_per_second": 0.457,
"step": 4000
}
],
"logging_steps": 1.0,
"max_steps": 13075,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.409350742875046e+18,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}