Files
Role_LLM_Cube_v1/trainer_state.json
ModelHub XC 5cc4441c3f 初始化项目,由ModelHub XC社区提供模型
Model: Lie24/Role_LLM_Cube_v1
Source: Original Platform
2026-04-28 18:19:05 +08:00

3340 lines
80 KiB
JSON

{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 10.0,
"eval_steps": 200,
"global_step": 470,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.02127659574468085,
"grad_norm": 19.769559521884688,
"learning_rate": 2.5000000000000004e-07,
"loss": 1.5136,
"step": 1
},
{
"epoch": 0.0425531914893617,
"grad_norm": 22.366909847966127,
"learning_rate": 5.000000000000001e-07,
"loss": 1.4656,
"step": 2
},
{
"epoch": 0.06382978723404255,
"grad_norm": 27.397989604017898,
"learning_rate": 7.5e-07,
"loss": 1.5124,
"step": 3
},
{
"epoch": 0.0851063829787234,
"grad_norm": 18.58567583552412,
"learning_rate": 1.0000000000000002e-06,
"loss": 1.5225,
"step": 4
},
{
"epoch": 0.10638297872340426,
"grad_norm": 18.165996038203307,
"learning_rate": 1.25e-06,
"loss": 1.5576,
"step": 5
},
{
"epoch": 0.1276595744680851,
"grad_norm": 13.681392535187342,
"learning_rate": 1.5e-06,
"loss": 1.4687,
"step": 6
},
{
"epoch": 0.14893617021276595,
"grad_norm": 14.999691223127021,
"learning_rate": 1.75e-06,
"loss": 1.3933,
"step": 7
},
{
"epoch": 0.1702127659574468,
"grad_norm": 9.360244090412072,
"learning_rate": 2.0000000000000003e-06,
"loss": 1.4721,
"step": 8
},
{
"epoch": 0.19148936170212766,
"grad_norm": 11.174023539278728,
"learning_rate": 2.25e-06,
"loss": 1.44,
"step": 9
},
{
"epoch": 0.2127659574468085,
"grad_norm": 9.866648780480253,
"learning_rate": 2.5e-06,
"loss": 1.2595,
"step": 10
},
{
"epoch": 0.23404255319148937,
"grad_norm": 9.512446051741868,
"learning_rate": 2.7500000000000004e-06,
"loss": 1.4442,
"step": 11
},
{
"epoch": 0.2553191489361702,
"grad_norm": 8.670274331402473,
"learning_rate": 3e-06,
"loss": 1.2739,
"step": 12
},
{
"epoch": 0.2765957446808511,
"grad_norm": 6.920817280642445,
"learning_rate": 3.2500000000000002e-06,
"loss": 1.276,
"step": 13
},
{
"epoch": 0.2978723404255319,
"grad_norm": 7.10414215318037,
"learning_rate": 3.5e-06,
"loss": 1.2421,
"step": 14
},
{
"epoch": 0.3191489361702128,
"grad_norm": 6.159725671936445,
"learning_rate": 3.7500000000000005e-06,
"loss": 1.239,
"step": 15
},
{
"epoch": 0.3404255319148936,
"grad_norm": 7.6873175205954825,
"learning_rate": 4.000000000000001e-06,
"loss": 1.2396,
"step": 16
},
{
"epoch": 0.3617021276595745,
"grad_norm": 6.528518611068642,
"learning_rate": 4.25e-06,
"loss": 1.1543,
"step": 17
},
{
"epoch": 0.3829787234042553,
"grad_norm": 6.77428307591927,
"learning_rate": 4.5e-06,
"loss": 1.1545,
"step": 18
},
{
"epoch": 0.40425531914893614,
"grad_norm": 6.417164213417,
"learning_rate": 4.75e-06,
"loss": 1.0959,
"step": 19
},
{
"epoch": 0.425531914893617,
"grad_norm": 6.180266176299509,
"learning_rate": 5e-06,
"loss": 1.1381,
"step": 20
},
{
"epoch": 0.44680851063829785,
"grad_norm": 6.238735118043302,
"learning_rate": 4.999939076763487e-06,
"loss": 0.9938,
"step": 21
},
{
"epoch": 0.46808510638297873,
"grad_norm": 5.503542713382496,
"learning_rate": 4.999756310023261e-06,
"loss": 1.0336,
"step": 22
},
{
"epoch": 0.48936170212765956,
"grad_norm": 5.7109281559241305,
"learning_rate": 4.999451708687114e-06,
"loss": 1.0293,
"step": 23
},
{
"epoch": 0.5106382978723404,
"grad_norm": 5.933533265865748,
"learning_rate": 4.999025287600886e-06,
"loss": 1.0838,
"step": 24
},
{
"epoch": 0.5319148936170213,
"grad_norm": 6.058382477727568,
"learning_rate": 4.99847706754774e-06,
"loss": 1.147,
"step": 25
},
{
"epoch": 0.5531914893617021,
"grad_norm": 5.039016064603872,
"learning_rate": 4.997807075247147e-06,
"loss": 1.0661,
"step": 26
},
{
"epoch": 0.574468085106383,
"grad_norm": 5.673653311719194,
"learning_rate": 4.9970153433535855e-06,
"loss": 1.1088,
"step": 27
},
{
"epoch": 0.5957446808510638,
"grad_norm": 6.772514217020031,
"learning_rate": 4.996101910454953e-06,
"loss": 1.0398,
"step": 28
},
{
"epoch": 0.6170212765957447,
"grad_norm": 7.261784604470834,
"learning_rate": 4.9950668210706795e-06,
"loss": 1.1197,
"step": 29
},
{
"epoch": 0.6382978723404256,
"grad_norm": 5.125305508590899,
"learning_rate": 4.993910125649561e-06,
"loss": 1.0096,
"step": 30
},
{
"epoch": 0.6595744680851063,
"grad_norm": 5.79320276945061,
"learning_rate": 4.992631880567301e-06,
"loss": 1.0418,
"step": 31
},
{
"epoch": 0.6808510638297872,
"grad_norm": 6.583723424976928,
"learning_rate": 4.9912321481237616e-06,
"loss": 1.1084,
"step": 32
},
{
"epoch": 0.7021276595744681,
"grad_norm": 5.166238541818618,
"learning_rate": 4.989710996539926e-06,
"loss": 1.0269,
"step": 33
},
{
"epoch": 0.723404255319149,
"grad_norm": 4.702597984819574,
"learning_rate": 4.988068499954578e-06,
"loss": 0.9362,
"step": 34
},
{
"epoch": 0.7446808510638298,
"grad_norm": 5.035553587199529,
"learning_rate": 4.986304738420684e-06,
"loss": 1.0027,
"step": 35
},
{
"epoch": 0.7659574468085106,
"grad_norm": 5.979469659293992,
"learning_rate": 4.984419797901491e-06,
"loss": 1.0197,
"step": 36
},
{
"epoch": 0.7872340425531915,
"grad_norm": 7.251693613828905,
"learning_rate": 4.9824137702663424e-06,
"loss": 1.0665,
"step": 37
},
{
"epoch": 0.8085106382978723,
"grad_norm": 5.336704026070371,
"learning_rate": 4.980286753286196e-06,
"loss": 1.0612,
"step": 38
},
{
"epoch": 0.8297872340425532,
"grad_norm": 4.572621845298497,
"learning_rate": 4.978038850628855e-06,
"loss": 0.891,
"step": 39
},
{
"epoch": 0.851063829787234,
"grad_norm": 5.901358668805014,
"learning_rate": 4.975670171853926e-06,
"loss": 1.003,
"step": 40
},
{
"epoch": 0.8723404255319149,
"grad_norm": 5.967637194891353,
"learning_rate": 4.973180832407471e-06,
"loss": 1.0412,
"step": 41
},
{
"epoch": 0.8936170212765957,
"grad_norm": 4.651650771697547,
"learning_rate": 4.970570953616383e-06,
"loss": 0.905,
"step": 42
},
{
"epoch": 0.9148936170212766,
"grad_norm": 4.40595275321189,
"learning_rate": 4.96784066268247e-06,
"loss": 0.9606,
"step": 43
},
{
"epoch": 0.9361702127659575,
"grad_norm": 4.297391018229843,
"learning_rate": 4.964990092676263e-06,
"loss": 0.912,
"step": 44
},
{
"epoch": 0.9574468085106383,
"grad_norm": 5.8520797403685565,
"learning_rate": 4.962019382530521e-06,
"loss": 0.99,
"step": 45
},
{
"epoch": 0.9787234042553191,
"grad_norm": 4.460582020214977,
"learning_rate": 4.958928677033465e-06,
"loss": 0.9338,
"step": 46
},
{
"epoch": 1.0,
"grad_norm": 5.108150405797771,
"learning_rate": 4.9557181268217225e-06,
"loss": 0.8841,
"step": 47
},
{
"epoch": 1.0212765957446808,
"grad_norm": 4.42640489710439,
"learning_rate": 4.9523878883729794e-06,
"loss": 0.6441,
"step": 48
},
{
"epoch": 1.0425531914893618,
"grad_norm": 4.783433777129786,
"learning_rate": 4.94893812399836e-06,
"loss": 0.7076,
"step": 49
},
{
"epoch": 1.0638297872340425,
"grad_norm": 4.461478664530093,
"learning_rate": 4.9453690018345144e-06,
"loss": 0.6507,
"step": 50
},
{
"epoch": 1.0851063829787233,
"grad_norm": 4.659068920568944,
"learning_rate": 4.9416806958354206e-06,
"loss": 0.6388,
"step": 51
},
{
"epoch": 1.1063829787234043,
"grad_norm": 5.326272476452319,
"learning_rate": 4.937873385763909e-06,
"loss": 0.6629,
"step": 52
},
{
"epoch": 1.127659574468085,
"grad_norm": 4.42497697146127,
"learning_rate": 4.933947257182901e-06,
"loss": 0.6969,
"step": 53
},
{
"epoch": 1.148936170212766,
"grad_norm": 4.342156315066868,
"learning_rate": 4.9299025014463665e-06,
"loss": 0.6389,
"step": 54
},
{
"epoch": 1.1702127659574468,
"grad_norm": 5.535251688604961,
"learning_rate": 4.925739315689991e-06,
"loss": 0.6549,
"step": 55
},
{
"epoch": 1.1914893617021276,
"grad_norm": 4.098290419568537,
"learning_rate": 4.921457902821578e-06,
"loss": 0.6581,
"step": 56
},
{
"epoch": 1.2127659574468086,
"grad_norm": 4.804191552305714,
"learning_rate": 4.917058471511149e-06,
"loss": 0.6574,
"step": 57
},
{
"epoch": 1.2340425531914894,
"grad_norm": 4.661998643228261,
"learning_rate": 4.912541236180779e-06,
"loss": 0.6206,
"step": 58
},
{
"epoch": 1.2553191489361701,
"grad_norm": 5.457436437790183,
"learning_rate": 4.907906416994146e-06,
"loss": 0.6949,
"step": 59
},
{
"epoch": 1.2765957446808511,
"grad_norm": 5.287463616155875,
"learning_rate": 4.903154239845798e-06,
"loss": 0.6171,
"step": 60
},
{
"epoch": 1.297872340425532,
"grad_norm": 6.534162370166582,
"learning_rate": 4.898284936350144e-06,
"loss": 0.6738,
"step": 61
},
{
"epoch": 1.3191489361702127,
"grad_norm": 5.14242301334199,
"learning_rate": 4.893298743830168e-06,
"loss": 0.71,
"step": 62
},
{
"epoch": 1.3404255319148937,
"grad_norm": 4.864653117677477,
"learning_rate": 4.888195905305859e-06,
"loss": 0.6353,
"step": 63
},
{
"epoch": 1.3617021276595744,
"grad_norm": 4.469502230256778,
"learning_rate": 4.882976669482368e-06,
"loss": 0.6301,
"step": 64
},
{
"epoch": 1.3829787234042552,
"grad_norm": 4.650400360608429,
"learning_rate": 4.8776412907378845e-06,
"loss": 0.6298,
"step": 65
},
{
"epoch": 1.4042553191489362,
"grad_norm": 4.75068049245489,
"learning_rate": 4.8721900291112415e-06,
"loss": 0.6132,
"step": 66
},
{
"epoch": 1.425531914893617,
"grad_norm": 5.406681668582622,
"learning_rate": 4.866623150289241e-06,
"loss": 0.6177,
"step": 67
},
{
"epoch": 1.4468085106382977,
"grad_norm": 4.689427420330228,
"learning_rate": 4.860940925593703e-06,
"loss": 0.601,
"step": 68
},
{
"epoch": 1.4680851063829787,
"grad_norm": 5.66362626435671,
"learning_rate": 4.855143631968242e-06,
"loss": 0.6987,
"step": 69
},
{
"epoch": 1.4893617021276595,
"grad_norm": 6.639330863120106,
"learning_rate": 4.849231551964771e-06,
"loss": 0.705,
"step": 70
},
{
"epoch": 1.5106382978723403,
"grad_norm": 5.15079243645691,
"learning_rate": 4.84320497372973e-06,
"loss": 0.6812,
"step": 71
},
{
"epoch": 1.5319148936170213,
"grad_norm": 5.318900788335571,
"learning_rate": 4.837064190990036e-06,
"loss": 0.6717,
"step": 72
},
{
"epoch": 1.5531914893617023,
"grad_norm": 4.39050844628742,
"learning_rate": 4.830809503038781e-06,
"loss": 0.5978,
"step": 73
},
{
"epoch": 1.574468085106383,
"grad_norm": 4.11981361882099,
"learning_rate": 4.824441214720629e-06,
"loss": 0.5789,
"step": 74
},
{
"epoch": 1.5957446808510638,
"grad_norm": 4.898106888331966,
"learning_rate": 4.817959636416969e-06,
"loss": 0.638,
"step": 75
},
{
"epoch": 1.6170212765957448,
"grad_norm": 4.144574853637901,
"learning_rate": 4.811365084030784e-06,
"loss": 0.6631,
"step": 76
},
{
"epoch": 1.6382978723404256,
"grad_norm": 4.236596267103849,
"learning_rate": 4.804657878971252e-06,
"loss": 0.5618,
"step": 77
},
{
"epoch": 1.6595744680851063,
"grad_norm": 4.772661867276851,
"learning_rate": 4.7978383481380865e-06,
"loss": 0.6064,
"step": 78
},
{
"epoch": 1.6808510638297873,
"grad_norm": 5.5176007762843335,
"learning_rate": 4.790906823905599e-06,
"loss": 0.6346,
"step": 79
},
{
"epoch": 1.702127659574468,
"grad_norm": 4.1371187239001825,
"learning_rate": 4.783863644106502e-06,
"loss": 0.598,
"step": 80
},
{
"epoch": 1.7234042553191489,
"grad_norm": 4.617726148114806,
"learning_rate": 4.776709152015443e-06,
"loss": 0.572,
"step": 81
},
{
"epoch": 1.7446808510638299,
"grad_norm": 4.389883430040871,
"learning_rate": 4.769443696332272e-06,
"loss": 0.6145,
"step": 82
},
{
"epoch": 1.7659574468085106,
"grad_norm": 5.027199601611735,
"learning_rate": 4.762067631165049e-06,
"loss": 0.6858,
"step": 83
},
{
"epoch": 1.7872340425531914,
"grad_norm": 4.990601804039334,
"learning_rate": 4.754581316012785e-06,
"loss": 0.561,
"step": 84
},
{
"epoch": 1.8085106382978724,
"grad_norm": 5.705018779334882,
"learning_rate": 4.746985115747918e-06,
"loss": 0.6323,
"step": 85
},
{
"epoch": 1.8297872340425532,
"grad_norm": 5.1505009350987265,
"learning_rate": 4.7392794005985324e-06,
"loss": 0.5952,
"step": 86
},
{
"epoch": 1.851063829787234,
"grad_norm": 5.16517962200903,
"learning_rate": 4.731464546130315e-06,
"loss": 0.6462,
"step": 87
},
{
"epoch": 1.872340425531915,
"grad_norm": 5.203202926098486,
"learning_rate": 4.723540933228245e-06,
"loss": 0.5949,
"step": 88
},
{
"epoch": 1.8936170212765957,
"grad_norm": 4.6706368589871134,
"learning_rate": 4.715508948078037e-06,
"loss": 0.5878,
"step": 89
},
{
"epoch": 1.9148936170212765,
"grad_norm": 4.351714864184194,
"learning_rate": 4.707368982147318e-06,
"loss": 0.6227,
"step": 90
},
{
"epoch": 1.9361702127659575,
"grad_norm": 5.11397965814302,
"learning_rate": 4.699121432166542e-06,
"loss": 0.6347,
"step": 91
},
{
"epoch": 1.9574468085106385,
"grad_norm": 5.875205347984843,
"learning_rate": 4.690766700109659e-06,
"loss": 0.5905,
"step": 92
},
{
"epoch": 1.978723404255319,
"grad_norm": 6.03539957826197,
"learning_rate": 4.682305193174524e-06,
"loss": 0.6894,
"step": 93
},
{
"epoch": 2.0,
"grad_norm": 3.8707434069223026,
"learning_rate": 4.673737323763048e-06,
"loss": 0.4095,
"step": 94
},
{
"epoch": 2.021276595744681,
"grad_norm": 4.967672001558827,
"learning_rate": 4.665063509461098e-06,
"loss": 0.3403,
"step": 95
},
{
"epoch": 2.0425531914893615,
"grad_norm": 3.5072435391900436,
"learning_rate": 4.656284173018144e-06,
"loss": 0.3736,
"step": 96
},
{
"epoch": 2.0638297872340425,
"grad_norm": 4.360376932941338,
"learning_rate": 4.6473997423266615e-06,
"loss": 0.4061,
"step": 97
},
{
"epoch": 2.0851063829787235,
"grad_norm": 4.224245961638772,
"learning_rate": 4.638410650401267e-06,
"loss": 0.3515,
"step": 98
},
{
"epoch": 2.106382978723404,
"grad_norm": 4.792845268520252,
"learning_rate": 4.62931733535762e-06,
"loss": 0.3432,
"step": 99
},
{
"epoch": 2.127659574468085,
"grad_norm": 4.9719261969262085,
"learning_rate": 4.620120240391065e-06,
"loss": 0.3495,
"step": 100
},
{
"epoch": 2.148936170212766,
"grad_norm": 3.9393952177630864,
"learning_rate": 4.610819813755038e-06,
"loss": 0.3505,
"step": 101
},
{
"epoch": 2.1702127659574466,
"grad_norm": 5.0676129989734084,
"learning_rate": 4.601416508739211e-06,
"loss": 0.345,
"step": 102
},
{
"epoch": 2.1914893617021276,
"grad_norm": 4.268873683189594,
"learning_rate": 4.591910783647405e-06,
"loss": 0.3411,
"step": 103
},
{
"epoch": 2.2127659574468086,
"grad_norm": 5.648372477585861,
"learning_rate": 4.582303101775249e-06,
"loss": 0.3414,
"step": 104
},
{
"epoch": 2.2340425531914896,
"grad_norm": 4.569109738179888,
"learning_rate": 4.572593931387604e-06,
"loss": 0.3107,
"step": 105
},
{
"epoch": 2.25531914893617,
"grad_norm": 5.986604177134095,
"learning_rate": 4.562783745695738e-06,
"loss": 0.358,
"step": 106
},
{
"epoch": 2.276595744680851,
"grad_norm": 4.593476704693552,
"learning_rate": 4.55287302283426e-06,
"loss": 0.3938,
"step": 107
},
{
"epoch": 2.297872340425532,
"grad_norm": 4.729132719197455,
"learning_rate": 4.542862245837821e-06,
"loss": 0.3445,
"step": 108
},
{
"epoch": 2.3191489361702127,
"grad_norm": 5.319389221723198,
"learning_rate": 4.5327519026175694e-06,
"loss": 0.3638,
"step": 109
},
{
"epoch": 2.3404255319148937,
"grad_norm": 4.9962223266535295,
"learning_rate": 4.522542485937369e-06,
"loss": 0.3151,
"step": 110
},
{
"epoch": 2.3617021276595747,
"grad_norm": 5.43349545021367,
"learning_rate": 4.512234493389785e-06,
"loss": 0.3386,
"step": 111
},
{
"epoch": 2.382978723404255,
"grad_norm": 4.228899687733491,
"learning_rate": 4.501828427371834e-06,
"loss": 0.3381,
"step": 112
},
{
"epoch": 2.404255319148936,
"grad_norm": 3.9364311500151214,
"learning_rate": 4.491324795060491e-06,
"loss": 0.3419,
"step": 113
},
{
"epoch": 2.425531914893617,
"grad_norm": 6.072764335330948,
"learning_rate": 4.4807241083879774e-06,
"loss": 0.3447,
"step": 114
},
{
"epoch": 2.4468085106382977,
"grad_norm": 4.690579387500509,
"learning_rate": 4.470026884016805e-06,
"loss": 0.3296,
"step": 115
},
{
"epoch": 2.4680851063829787,
"grad_norm": 4.772781283406453,
"learning_rate": 4.4592336433146e-06,
"loss": 0.3112,
"step": 116
},
{
"epoch": 2.4893617021276597,
"grad_norm": 4.222608771767499,
"learning_rate": 4.448344912328686e-06,
"loss": 0.3602,
"step": 117
},
{
"epoch": 2.5106382978723403,
"grad_norm": 4.806194387046772,
"learning_rate": 4.437361221760449e-06,
"loss": 0.3433,
"step": 118
},
{
"epoch": 2.5319148936170213,
"grad_norm": 4.610773667593825,
"learning_rate": 4.426283106939474e-06,
"loss": 0.3423,
"step": 119
},
{
"epoch": 2.5531914893617023,
"grad_norm": 3.9378113503820114,
"learning_rate": 4.415111107797445e-06,
"loss": 0.335,
"step": 120
},
{
"epoch": 2.574468085106383,
"grad_norm": 4.262317442324841,
"learning_rate": 4.403845768841842e-06,
"loss": 0.2964,
"step": 121
},
{
"epoch": 2.595744680851064,
"grad_norm": 4.586717288805483,
"learning_rate": 4.3924876391293915e-06,
"loss": 0.3503,
"step": 122
},
{
"epoch": 2.617021276595745,
"grad_norm": 4.420522314781192,
"learning_rate": 4.381037272239311e-06,
"loss": 0.348,
"step": 123
},
{
"epoch": 2.6382978723404253,
"grad_norm": 3.6908626258947694,
"learning_rate": 4.36949522624633e-06,
"loss": 0.3473,
"step": 124
},
{
"epoch": 2.6595744680851063,
"grad_norm": 4.276095484034931,
"learning_rate": 4.357862063693486e-06,
"loss": 0.3734,
"step": 125
},
{
"epoch": 2.6808510638297873,
"grad_norm": 4.16461417799402,
"learning_rate": 4.346138351564711e-06,
"loss": 0.3563,
"step": 126
},
{
"epoch": 2.702127659574468,
"grad_norm": 4.106373804902514,
"learning_rate": 4.334324661257191e-06,
"loss": 0.2954,
"step": 127
},
{
"epoch": 2.723404255319149,
"grad_norm": 4.113950324587281,
"learning_rate": 4.322421568553529e-06,
"loss": 0.3443,
"step": 128
},
{
"epoch": 2.74468085106383,
"grad_norm": 4.190826822795899,
"learning_rate": 4.3104296535936695e-06,
"loss": 0.3395,
"step": 129
},
{
"epoch": 2.7659574468085104,
"grad_norm": 4.9676092124733575,
"learning_rate": 4.2983495008466285e-06,
"loss": 0.3789,
"step": 130
},
{
"epoch": 2.7872340425531914,
"grad_norm": 4.175058791875455,
"learning_rate": 4.286181699082008e-06,
"loss": 0.3388,
"step": 131
},
{
"epoch": 2.8085106382978724,
"grad_norm": 5.475615243400478,
"learning_rate": 4.273926841341303e-06,
"loss": 0.3614,
"step": 132
},
{
"epoch": 2.829787234042553,
"grad_norm": 4.132179706281526,
"learning_rate": 4.261585524908987e-06,
"loss": 0.3031,
"step": 133
},
{
"epoch": 2.851063829787234,
"grad_norm": 4.52680856181224,
"learning_rate": 4.249158351283414e-06,
"loss": 0.3202,
"step": 134
},
{
"epoch": 2.872340425531915,
"grad_norm": 4.4158856257333206,
"learning_rate": 4.236645926147493e-06,
"loss": 0.3817,
"step": 135
},
{
"epoch": 2.8936170212765955,
"grad_norm": 4.131675629353336,
"learning_rate": 4.224048859339175e-06,
"loss": 0.325,
"step": 136
},
{
"epoch": 2.9148936170212765,
"grad_norm": 4.632609969168001,
"learning_rate": 4.211367764821722e-06,
"loss": 0.339,
"step": 137
},
{
"epoch": 2.9361702127659575,
"grad_norm": 4.238399617644217,
"learning_rate": 4.198603260653792e-06,
"loss": 0.3803,
"step": 138
},
{
"epoch": 2.9574468085106385,
"grad_norm": 4.990761801935908,
"learning_rate": 4.185755968959308e-06,
"loss": 0.3447,
"step": 139
},
{
"epoch": 2.978723404255319,
"grad_norm": 4.548241919258143,
"learning_rate": 4.172826515897146e-06,
"loss": 0.3518,
"step": 140
},
{
"epoch": 3.0,
"grad_norm": 3.430910087137401,
"learning_rate": 4.159815531630604e-06,
"loss": 0.2304,
"step": 141
},
{
"epoch": 3.021276595744681,
"grad_norm": 3.668126527915725,
"learning_rate": 4.146723650296701e-06,
"loss": 0.2235,
"step": 142
},
{
"epoch": 3.0425531914893615,
"grad_norm": 2.8667931520747327,
"learning_rate": 4.133551509975264e-06,
"loss": 0.1883,
"step": 143
},
{
"epoch": 3.0638297872340425,
"grad_norm": 3.5239577403765106,
"learning_rate": 4.120299752657828e-06,
"loss": 0.2217,
"step": 144
},
{
"epoch": 3.0851063829787235,
"grad_norm": 3.6921907938710308,
"learning_rate": 4.106969024216348e-06,
"loss": 0.2049,
"step": 145
},
{
"epoch": 3.106382978723404,
"grad_norm": 3.6392973250804577,
"learning_rate": 4.093559974371725e-06,
"loss": 0.1831,
"step": 146
},
{
"epoch": 3.127659574468085,
"grad_norm": 3.7365922708313133,
"learning_rate": 4.080073256662128e-06,
"loss": 0.198,
"step": 147
},
{
"epoch": 3.148936170212766,
"grad_norm": 4.42100805526352,
"learning_rate": 4.066509528411151e-06,
"loss": 0.1593,
"step": 148
},
{
"epoch": 3.1702127659574466,
"grad_norm": 3.701129708479829,
"learning_rate": 4.052869450695776e-06,
"loss": 0.1542,
"step": 149
},
{
"epoch": 3.1914893617021276,
"grad_norm": 3.9620190917388793,
"learning_rate": 4.039153688314146e-06,
"loss": 0.1673,
"step": 150
},
{
"epoch": 3.2127659574468086,
"grad_norm": 4.596659093022736,
"learning_rate": 4.02536290975317e-06,
"loss": 0.1781,
"step": 151
},
{
"epoch": 3.2340425531914896,
"grad_norm": 4.041018040517897,
"learning_rate": 4.011497787155938e-06,
"loss": 0.204,
"step": 152
},
{
"epoch": 3.25531914893617,
"grad_norm": 4.4750076668231324,
"learning_rate": 3.997558996288965e-06,
"loss": 0.2042,
"step": 153
},
{
"epoch": 3.276595744680851,
"grad_norm": 4.328565873011614,
"learning_rate": 3.983547216509254e-06,
"loss": 0.206,
"step": 154
},
{
"epoch": 3.297872340425532,
"grad_norm": 4.950916510385211,
"learning_rate": 3.969463130731183e-06,
"loss": 0.1735,
"step": 155
},
{
"epoch": 3.3191489361702127,
"grad_norm": 3.6634515919781356,
"learning_rate": 3.955307425393224e-06,
"loss": 0.2152,
"step": 156
},
{
"epoch": 3.3404255319148937,
"grad_norm": 3.5183557309327127,
"learning_rate": 3.941080790424483e-06,
"loss": 0.1788,
"step": 157
},
{
"epoch": 3.3617021276595747,
"grad_norm": 4.2312437308996245,
"learning_rate": 3.92678391921108e-06,
"loss": 0.1817,
"step": 158
},
{
"epoch": 3.382978723404255,
"grad_norm": 3.4224234758888654,
"learning_rate": 3.912417508562345e-06,
"loss": 0.1577,
"step": 159
},
{
"epoch": 3.404255319148936,
"grad_norm": 3.7251058715223597,
"learning_rate": 3.897982258676867e-06,
"loss": 0.2067,
"step": 160
},
{
"epoch": 3.425531914893617,
"grad_norm": 3.123331948611014,
"learning_rate": 3.88347887310836e-06,
"loss": 0.1867,
"step": 161
},
{
"epoch": 3.4468085106382977,
"grad_norm": 3.040022180599425,
"learning_rate": 3.868908058731376e-06,
"loss": 0.1488,
"step": 162
},
{
"epoch": 3.4680851063829787,
"grad_norm": 3.936251625166522,
"learning_rate": 3.85427052570685e-06,
"loss": 0.175,
"step": 163
},
{
"epoch": 3.4893617021276597,
"grad_norm": 3.35819244249543,
"learning_rate": 3.839566987447492e-06,
"loss": 0.1992,
"step": 164
},
{
"epoch": 3.5106382978723403,
"grad_norm": 3.8098483118828335,
"learning_rate": 3.824798160583012e-06,
"loss": 0.1677,
"step": 165
},
{
"epoch": 3.5319148936170213,
"grad_norm": 4.08720146453336,
"learning_rate": 3.8099647649251984e-06,
"loss": 0.1975,
"step": 166
},
{
"epoch": 3.5531914893617023,
"grad_norm": 3.546201338791242,
"learning_rate": 3.795067523432826e-06,
"loss": 0.1643,
"step": 167
},
{
"epoch": 3.574468085106383,
"grad_norm": 4.022986409946214,
"learning_rate": 3.780107162176429e-06,
"loss": 0.2025,
"step": 168
},
{
"epoch": 3.595744680851064,
"grad_norm": 4.116301951633153,
"learning_rate": 3.7650844103029093e-06,
"loss": 0.2063,
"step": 169
},
{
"epoch": 3.617021276595745,
"grad_norm": 5.498041562915274,
"learning_rate": 3.7500000000000005e-06,
"loss": 0.1969,
"step": 170
},
{
"epoch": 3.6382978723404253,
"grad_norm": 3.571990562222316,
"learning_rate": 3.7348546664605777e-06,
"loss": 0.1946,
"step": 171
},
{
"epoch": 3.6595744680851063,
"grad_norm": 4.110746067790737,
"learning_rate": 3.7196491478468322e-06,
"loss": 0.192,
"step": 172
},
{
"epoch": 3.6808510638297873,
"grad_norm": 3.9747616004152984,
"learning_rate": 3.7043841852542884e-06,
"loss": 0.2076,
"step": 173
},
{
"epoch": 3.702127659574468,
"grad_norm": 4.049886537167864,
"learning_rate": 3.689060522675689e-06,
"loss": 0.1885,
"step": 174
},
{
"epoch": 3.723404255319149,
"grad_norm": 3.7605106851151593,
"learning_rate": 3.6736789069647273e-06,
"loss": 0.1819,
"step": 175
},
{
"epoch": 3.74468085106383,
"grad_norm": 4.293344674761706,
"learning_rate": 3.658240087799655e-06,
"loss": 0.1782,
"step": 176
},
{
"epoch": 3.7659574468085104,
"grad_norm": 3.094426165785156,
"learning_rate": 3.642744817646736e-06,
"loss": 0.1562,
"step": 177
},
{
"epoch": 3.7872340425531914,
"grad_norm": 4.471233004565249,
"learning_rate": 3.627193851723577e-06,
"loss": 0.1538,
"step": 178
},
{
"epoch": 3.8085106382978724,
"grad_norm": 3.8346276158744823,
"learning_rate": 3.611587947962319e-06,
"loss": 0.2037,
"step": 179
},
{
"epoch": 3.829787234042553,
"grad_norm": 4.6309266206267745,
"learning_rate": 3.595927866972694e-06,
"loss": 0.1706,
"step": 180
},
{
"epoch": 3.851063829787234,
"grad_norm": 3.574650150496055,
"learning_rate": 3.5802143720049565e-06,
"loss": 0.1793,
"step": 181
},
{
"epoch": 3.872340425531915,
"grad_norm": 4.104316685959246,
"learning_rate": 3.564448228912682e-06,
"loss": 0.1997,
"step": 182
},
{
"epoch": 3.8936170212765955,
"grad_norm": 3.7077174376752815,
"learning_rate": 3.5486302061154433e-06,
"loss": 0.2397,
"step": 183
},
{
"epoch": 3.9148936170212765,
"grad_norm": 3.8535618259715863,
"learning_rate": 3.532761074561355e-06,
"loss": 0.1724,
"step": 184
},
{
"epoch": 3.9361702127659575,
"grad_norm": 4.153454591162772,
"learning_rate": 3.516841607689501e-06,
"loss": 0.1879,
"step": 185
},
{
"epoch": 3.9574468085106385,
"grad_norm": 3.2186742898572254,
"learning_rate": 3.5008725813922383e-06,
"loss": 0.1793,
"step": 186
},
{
"epoch": 3.978723404255319,
"grad_norm": 3.8852555348009066,
"learning_rate": 3.4848547739773782e-06,
"loss": 0.2205,
"step": 187
},
{
"epoch": 4.0,
"grad_norm": 3.1773700849249984,
"learning_rate": 3.4687889661302577e-06,
"loss": 0.1375,
"step": 188
},
{
"epoch": 4.0212765957446805,
"grad_norm": 2.909112047920458,
"learning_rate": 3.452675940875686e-06,
"loss": 0.1081,
"step": 189
},
{
"epoch": 4.042553191489362,
"grad_norm": 2.5683738553439794,
"learning_rate": 3.436516483539781e-06,
"loss": 0.0953,
"step": 190
},
{
"epoch": 4.0638297872340425,
"grad_norm": 2.9198721500124845,
"learning_rate": 3.4203113817116955e-06,
"loss": 0.1053,
"step": 191
},
{
"epoch": 4.085106382978723,
"grad_norm": 2.24010493827577,
"learning_rate": 3.4040614252052305e-06,
"loss": 0.0927,
"step": 192
},
{
"epoch": 4.1063829787234045,
"grad_norm": 2.878670182658427,
"learning_rate": 3.387767406020343e-06,
"loss": 0.103,
"step": 193
},
{
"epoch": 4.127659574468085,
"grad_norm": 3.1788187649976978,
"learning_rate": 3.3714301183045382e-06,
"loss": 0.1118,
"step": 194
},
{
"epoch": 4.148936170212766,
"grad_norm": 3.3067413198545634,
"learning_rate": 3.3550503583141726e-06,
"loss": 0.1099,
"step": 195
},
{
"epoch": 4.170212765957447,
"grad_norm": 2.319050087976373,
"learning_rate": 3.338628924375638e-06,
"loss": 0.0898,
"step": 196
},
{
"epoch": 4.191489361702128,
"grad_norm": 2.6474547453865895,
"learning_rate": 3.3221666168464584e-06,
"loss": 0.0915,
"step": 197
},
{
"epoch": 4.212765957446808,
"grad_norm": 3.2898592433873217,
"learning_rate": 3.3056642380762783e-06,
"loss": 0.076,
"step": 198
},
{
"epoch": 4.23404255319149,
"grad_norm": 3.7084220696956685,
"learning_rate": 3.2891225923677565e-06,
"loss": 0.0832,
"step": 199
},
{
"epoch": 4.25531914893617,
"grad_norm": 3.853474111497437,
"learning_rate": 3.272542485937369e-06,
"loss": 0.1113,
"step": 200
},
{
"epoch": 4.25531914893617,
"eval_loss": 1.3551427125930786,
"eval_runtime": 1.3449,
"eval_samples_per_second": 17.102,
"eval_steps_per_second": 0.744,
"step": 200
},
{
"epoch": 4.276595744680851,
"grad_norm": 3.1480732576505686,
"learning_rate": 3.2559247268761117e-06,
"loss": 0.0894,
"step": 201
},
{
"epoch": 4.297872340425532,
"grad_norm": 4.224708808728799,
"learning_rate": 3.2392701251101172e-06,
"loss": 0.1037,
"step": 202
},
{
"epoch": 4.319148936170213,
"grad_norm": 3.127550241031144,
"learning_rate": 3.222579492361179e-06,
"loss": 0.1193,
"step": 203
},
{
"epoch": 4.340425531914893,
"grad_norm": 3.358278204814404,
"learning_rate": 3.205853642107192e-06,
"loss": 0.1088,
"step": 204
},
{
"epoch": 4.361702127659575,
"grad_norm": 3.5723872992052668,
"learning_rate": 3.189093389542498e-06,
"loss": 0.1156,
"step": 205
},
{
"epoch": 4.382978723404255,
"grad_norm": 3.3910498011991863,
"learning_rate": 3.1722995515381644e-06,
"loss": 0.0813,
"step": 206
},
{
"epoch": 4.404255319148936,
"grad_norm": 3.987898875106463,
"learning_rate": 3.155472946602162e-06,
"loss": 0.1063,
"step": 207
},
{
"epoch": 4.425531914893617,
"grad_norm": 3.0288381078835416,
"learning_rate": 3.1386143948394764e-06,
"loss": 0.1089,
"step": 208
},
{
"epoch": 4.446808510638298,
"grad_norm": 3.3307094726424635,
"learning_rate": 3.121724717912138e-06,
"loss": 0.1142,
"step": 209
},
{
"epoch": 4.468085106382979,
"grad_norm": 2.9848172012011664,
"learning_rate": 3.1048047389991693e-06,
"loss": 0.1157,
"step": 210
},
{
"epoch": 4.48936170212766,
"grad_norm": 2.978438802813703,
"learning_rate": 3.087855282756475e-06,
"loss": 0.0942,
"step": 211
},
{
"epoch": 4.51063829787234,
"grad_norm": 2.8670301514814565,
"learning_rate": 3.0708771752766397e-06,
"loss": 0.0876,
"step": 212
},
{
"epoch": 4.531914893617021,
"grad_norm": 3.0160259648937626,
"learning_rate": 3.053871244048669e-06,
"loss": 0.1057,
"step": 213
},
{
"epoch": 4.553191489361702,
"grad_norm": 3.552730667163276,
"learning_rate": 3.0368383179176584e-06,
"loss": 0.0681,
"step": 214
},
{
"epoch": 4.574468085106383,
"grad_norm": 3.0671796344169744,
"learning_rate": 3.019779227044398e-06,
"loss": 0.1052,
"step": 215
},
{
"epoch": 4.595744680851064,
"grad_norm": 3.4917736614570343,
"learning_rate": 3.002694802864912e-06,
"loss": 0.1131,
"step": 216
},
{
"epoch": 4.617021276595745,
"grad_norm": 2.997068423954694,
"learning_rate": 2.98558587804993e-06,
"loss": 0.1249,
"step": 217
},
{
"epoch": 4.638297872340425,
"grad_norm": 2.945598337302224,
"learning_rate": 2.9684532864643123e-06,
"loss": 0.0916,
"step": 218
},
{
"epoch": 4.659574468085106,
"grad_norm": 2.5534208875346494,
"learning_rate": 2.9512978631264006e-06,
"loss": 0.0834,
"step": 219
},
{
"epoch": 4.680851063829787,
"grad_norm": 2.3947639549989144,
"learning_rate": 2.9341204441673267e-06,
"loss": 0.0974,
"step": 220
},
{
"epoch": 4.702127659574468,
"grad_norm": 3.3251173109542074,
"learning_rate": 2.9169218667902562e-06,
"loss": 0.1011,
"step": 221
},
{
"epoch": 4.723404255319149,
"grad_norm": 3.059268693430171,
"learning_rate": 2.8997029692295875e-06,
"loss": 0.0821,
"step": 222
},
{
"epoch": 4.74468085106383,
"grad_norm": 3.5597698182253965,
"learning_rate": 2.8824645907100957e-06,
"loss": 0.0856,
"step": 223
},
{
"epoch": 4.76595744680851,
"grad_norm": 3.3566578614960423,
"learning_rate": 2.8652075714060296e-06,
"loss": 0.132,
"step": 224
},
{
"epoch": 4.787234042553192,
"grad_norm": 3.470592088594595,
"learning_rate": 2.847932752400164e-06,
"loss": 0.0958,
"step": 225
},
{
"epoch": 4.808510638297872,
"grad_norm": 2.911346797843613,
"learning_rate": 2.8306409756428067e-06,
"loss": 0.0894,
"step": 226
},
{
"epoch": 4.829787234042553,
"grad_norm": 4.613111358505509,
"learning_rate": 2.813333083910761e-06,
"loss": 0.1166,
"step": 227
},
{
"epoch": 4.851063829787234,
"grad_norm": 3.015962228404618,
"learning_rate": 2.7960099207662535e-06,
"loss": 0.1155,
"step": 228
},
{
"epoch": 4.872340425531915,
"grad_norm": 2.987399829854267,
"learning_rate": 2.778672330515814e-06,
"loss": 0.1104,
"step": 229
},
{
"epoch": 4.8936170212765955,
"grad_norm": 2.829987084084442,
"learning_rate": 2.761321158169134e-06,
"loss": 0.1066,
"step": 230
},
{
"epoch": 4.914893617021277,
"grad_norm": 3.708679514606222,
"learning_rate": 2.743957249397874e-06,
"loss": 0.1048,
"step": 231
},
{
"epoch": 4.9361702127659575,
"grad_norm": 2.749591872522463,
"learning_rate": 2.726581450494451e-06,
"loss": 0.1159,
"step": 232
},
{
"epoch": 4.957446808510638,
"grad_norm": 3.776661504783798,
"learning_rate": 2.70919460833079e-06,
"loss": 0.1167,
"step": 233
},
{
"epoch": 4.9787234042553195,
"grad_norm": 3.091697194785843,
"learning_rate": 2.6917975703170466e-06,
"loss": 0.1174,
"step": 234
},
{
"epoch": 5.0,
"grad_norm": 2.8077659070563192,
"learning_rate": 2.6743911843603134e-06,
"loss": 0.0737,
"step": 235
},
{
"epoch": 5.0212765957446805,
"grad_norm": 1.9864860542138973,
"learning_rate": 2.6569762988232838e-06,
"loss": 0.0498,
"step": 236
},
{
"epoch": 5.042553191489362,
"grad_norm": 2.080148938371656,
"learning_rate": 2.63955376248291e-06,
"loss": 0.0588,
"step": 237
},
{
"epoch": 5.0638297872340425,
"grad_norm": 2.4271830087344215,
"learning_rate": 2.6221244244890336e-06,
"loss": 0.04,
"step": 238
},
{
"epoch": 5.085106382978723,
"grad_norm": 2.182065583755821,
"learning_rate": 2.604689134322999e-06,
"loss": 0.0635,
"step": 239
},
{
"epoch": 5.1063829787234045,
"grad_norm": 2.3865912853026727,
"learning_rate": 2.587248741756253e-06,
"loss": 0.0602,
"step": 240
},
{
"epoch": 5.127659574468085,
"grad_norm": 2.3131711827099832,
"learning_rate": 2.569804096808923e-06,
"loss": 0.0709,
"step": 241
},
{
"epoch": 5.148936170212766,
"grad_norm": 2.3050528514219004,
"learning_rate": 2.5523560497083927e-06,
"loss": 0.0547,
"step": 242
},
{
"epoch": 5.170212765957447,
"grad_norm": 2.124162166492525,
"learning_rate": 2.5349054508478636e-06,
"loss": 0.046,
"step": 243
},
{
"epoch": 5.191489361702128,
"grad_norm": 1.9773286780066965,
"learning_rate": 2.517453150744904e-06,
"loss": 0.0432,
"step": 244
},
{
"epoch": 5.212765957446808,
"grad_norm": 3.116678592316818,
"learning_rate": 2.5e-06,
"loss": 0.0634,
"step": 245
},
{
"epoch": 5.23404255319149,
"grad_norm": 2.504016738413573,
"learning_rate": 2.482546849255096e-06,
"loss": 0.0589,
"step": 246
},
{
"epoch": 5.25531914893617,
"grad_norm": 3.505324999262586,
"learning_rate": 2.4650945491521372e-06,
"loss": 0.0547,
"step": 247
},
{
"epoch": 5.276595744680851,
"grad_norm": 2.2697348908750974,
"learning_rate": 2.447643950291608e-06,
"loss": 0.0608,
"step": 248
},
{
"epoch": 5.297872340425532,
"grad_norm": 2.8775304018384986,
"learning_rate": 2.4301959031910785e-06,
"loss": 0.0765,
"step": 249
},
{
"epoch": 5.319148936170213,
"grad_norm": 2.4093465444498614,
"learning_rate": 2.4127512582437486e-06,
"loss": 0.0583,
"step": 250
},
{
"epoch": 5.340425531914893,
"grad_norm": 2.365667376680801,
"learning_rate": 2.3953108656770018e-06,
"loss": 0.0629,
"step": 251
},
{
"epoch": 5.361702127659575,
"grad_norm": 2.713559880628167,
"learning_rate": 2.377875575510967e-06,
"loss": 0.0669,
"step": 252
},
{
"epoch": 5.382978723404255,
"grad_norm": 2.999000225951145,
"learning_rate": 2.3604462375170905e-06,
"loss": 0.0617,
"step": 253
},
{
"epoch": 5.404255319148936,
"grad_norm": 2.6164170571843512,
"learning_rate": 2.3430237011767166e-06,
"loss": 0.0582,
"step": 254
},
{
"epoch": 5.425531914893617,
"grad_norm": 2.4039871330574796,
"learning_rate": 2.325608815639687e-06,
"loss": 0.0578,
"step": 255
},
{
"epoch": 5.446808510638298,
"grad_norm": 2.97747740854911,
"learning_rate": 2.3082024296829538e-06,
"loss": 0.0659,
"step": 256
},
{
"epoch": 5.468085106382979,
"grad_norm": 2.336759542261552,
"learning_rate": 2.290805391669212e-06,
"loss": 0.0548,
"step": 257
},
{
"epoch": 5.48936170212766,
"grad_norm": 3.1670650790685846,
"learning_rate": 2.2734185495055503e-06,
"loss": 0.0585,
"step": 258
},
{
"epoch": 5.51063829787234,
"grad_norm": 3.2034707124777952,
"learning_rate": 2.256042750602127e-06,
"loss": 0.0593,
"step": 259
},
{
"epoch": 5.531914893617021,
"grad_norm": 2.3348070175379503,
"learning_rate": 2.238678841830867e-06,
"loss": 0.0547,
"step": 260
},
{
"epoch": 5.553191489361702,
"grad_norm": 2.608427512535968,
"learning_rate": 2.2213276694841866e-06,
"loss": 0.0643,
"step": 261
},
{
"epoch": 5.574468085106383,
"grad_norm": 2.4848496472118495,
"learning_rate": 2.2039900792337477e-06,
"loss": 0.0608,
"step": 262
},
{
"epoch": 5.595744680851064,
"grad_norm": 1.886510569911626,
"learning_rate": 2.186666916089239e-06,
"loss": 0.0599,
"step": 263
},
{
"epoch": 5.617021276595745,
"grad_norm": 2.8635289633291547,
"learning_rate": 2.1693590243571937e-06,
"loss": 0.0515,
"step": 264
},
{
"epoch": 5.638297872340425,
"grad_norm": 2.854324373551202,
"learning_rate": 2.1520672475998374e-06,
"loss": 0.0554,
"step": 265
},
{
"epoch": 5.659574468085106,
"grad_norm": 1.8544197687406436,
"learning_rate": 2.134792428593971e-06,
"loss": 0.0421,
"step": 266
},
{
"epoch": 5.680851063829787,
"grad_norm": 2.6205983731152322,
"learning_rate": 2.117535409289905e-06,
"loss": 0.0457,
"step": 267
},
{
"epoch": 5.702127659574468,
"grad_norm": 1.7630228804740633,
"learning_rate": 2.1002970307704134e-06,
"loss": 0.0369,
"step": 268
},
{
"epoch": 5.723404255319149,
"grad_norm": 2.638068560108755,
"learning_rate": 2.0830781332097446e-06,
"loss": 0.0583,
"step": 269
},
{
"epoch": 5.74468085106383,
"grad_norm": 3.1753027345645646,
"learning_rate": 2.0658795558326745e-06,
"loss": 0.0674,
"step": 270
},
{
"epoch": 5.76595744680851,
"grad_norm": 2.2642667997058257,
"learning_rate": 2.0487021368736002e-06,
"loss": 0.064,
"step": 271
},
{
"epoch": 5.787234042553192,
"grad_norm": 2.559426313332899,
"learning_rate": 2.031546713535688e-06,
"loss": 0.0549,
"step": 272
},
{
"epoch": 5.808510638297872,
"grad_norm": 2.878855059930482,
"learning_rate": 2.0144141219500707e-06,
"loss": 0.0559,
"step": 273
},
{
"epoch": 5.829787234042553,
"grad_norm": 2.283778113411308,
"learning_rate": 1.997305197135089e-06,
"loss": 0.0548,
"step": 274
},
{
"epoch": 5.851063829787234,
"grad_norm": 2.4192051614394003,
"learning_rate": 1.9802207729556023e-06,
"loss": 0.0502,
"step": 275
},
{
"epoch": 5.872340425531915,
"grad_norm": 2.322479824517103,
"learning_rate": 1.963161682082342e-06,
"loss": 0.064,
"step": 276
},
{
"epoch": 5.8936170212765955,
"grad_norm": 2.4593833670456005,
"learning_rate": 1.946128755951332e-06,
"loss": 0.058,
"step": 277
},
{
"epoch": 5.914893617021277,
"grad_norm": 2.5251545274407907,
"learning_rate": 1.9291228247233607e-06,
"loss": 0.0522,
"step": 278
},
{
"epoch": 5.9361702127659575,
"grad_norm": 1.8104094730869373,
"learning_rate": 1.912144717243525e-06,
"loss": 0.0296,
"step": 279
},
{
"epoch": 5.957446808510638,
"grad_norm": 2.689146388919388,
"learning_rate": 1.895195261000831e-06,
"loss": 0.0407,
"step": 280
},
{
"epoch": 5.9787234042553195,
"grad_norm": 1.8669924342723785,
"learning_rate": 1.8782752820878636e-06,
"loss": 0.0495,
"step": 281
},
{
"epoch": 6.0,
"grad_norm": 2.271539062015242,
"learning_rate": 1.8613856051605242e-06,
"loss": 0.0455,
"step": 282
},
{
"epoch": 6.0212765957446805,
"grad_norm": 1.525504212144237,
"learning_rate": 1.8445270533978387e-06,
"loss": 0.0377,
"step": 283
},
{
"epoch": 6.042553191489362,
"grad_norm": 1.0859867252268387,
"learning_rate": 1.827700448461836e-06,
"loss": 0.0221,
"step": 284
},
{
"epoch": 6.0638297872340425,
"grad_norm": 1.6689218206446002,
"learning_rate": 1.8109066104575023e-06,
"loss": 0.0237,
"step": 285
},
{
"epoch": 6.085106382978723,
"grad_norm": 1.7091034732477,
"learning_rate": 1.7941463578928088e-06,
"loss": 0.0316,
"step": 286
},
{
"epoch": 6.1063829787234045,
"grad_norm": 1.473946493341099,
"learning_rate": 1.7774205076388207e-06,
"loss": 0.0301,
"step": 287
},
{
"epoch": 6.127659574468085,
"grad_norm": 1.9210720976922275,
"learning_rate": 1.7607298748898844e-06,
"loss": 0.0241,
"step": 288
},
{
"epoch": 6.148936170212766,
"grad_norm": 2.2803503758262504,
"learning_rate": 1.744075273123889e-06,
"loss": 0.0454,
"step": 289
},
{
"epoch": 6.170212765957447,
"grad_norm": 2.024197465049683,
"learning_rate": 1.7274575140626318e-06,
"loss": 0.0362,
"step": 290
},
{
"epoch": 6.191489361702128,
"grad_norm": 1.84667704080016,
"learning_rate": 1.7108774076322443e-06,
"loss": 0.0403,
"step": 291
},
{
"epoch": 6.212765957446808,
"grad_norm": 1.695502301672247,
"learning_rate": 1.6943357619237227e-06,
"loss": 0.0306,
"step": 292
},
{
"epoch": 6.23404255319149,
"grad_norm": 1.6806451272904126,
"learning_rate": 1.677833383153542e-06,
"loss": 0.0284,
"step": 293
},
{
"epoch": 6.25531914893617,
"grad_norm": 2.9118161159346854,
"learning_rate": 1.661371075624363e-06,
"loss": 0.0256,
"step": 294
},
{
"epoch": 6.276595744680851,
"grad_norm": 1.8173030910949344,
"learning_rate": 1.6449496416858285e-06,
"loss": 0.0284,
"step": 295
},
{
"epoch": 6.297872340425532,
"grad_norm": 2.223540017311328,
"learning_rate": 1.6285698816954626e-06,
"loss": 0.0443,
"step": 296
},
{
"epoch": 6.319148936170213,
"grad_norm": 1.4963288286875704,
"learning_rate": 1.612232593979658e-06,
"loss": 0.0367,
"step": 297
},
{
"epoch": 6.340425531914893,
"grad_norm": 1.3236418660083258,
"learning_rate": 1.5959385747947697e-06,
"loss": 0.0168,
"step": 298
},
{
"epoch": 6.361702127659575,
"grad_norm": 2.1082580928006704,
"learning_rate": 1.5796886182883053e-06,
"loss": 0.0384,
"step": 299
},
{
"epoch": 6.382978723404255,
"grad_norm": 2.0461518407784665,
"learning_rate": 1.56348351646022e-06,
"loss": 0.0348,
"step": 300
},
{
"epoch": 6.404255319148936,
"grad_norm": 1.9027154330548208,
"learning_rate": 1.547324059124315e-06,
"loss": 0.0409,
"step": 301
},
{
"epoch": 6.425531914893617,
"grad_norm": 1.8430627981587546,
"learning_rate": 1.5312110338697427e-06,
"loss": 0.0336,
"step": 302
},
{
"epoch": 6.446808510638298,
"grad_norm": 2.013051690553332,
"learning_rate": 1.5151452260226224e-06,
"loss": 0.0361,
"step": 303
},
{
"epoch": 6.468085106382979,
"grad_norm": 2.1207015485047753,
"learning_rate": 1.4991274186077632e-06,
"loss": 0.0405,
"step": 304
},
{
"epoch": 6.48936170212766,
"grad_norm": 1.3483418905294124,
"learning_rate": 1.4831583923105e-06,
"loss": 0.0264,
"step": 305
},
{
"epoch": 6.51063829787234,
"grad_norm": 1.2898003351353924,
"learning_rate": 1.467238925438646e-06,
"loss": 0.0326,
"step": 306
},
{
"epoch": 6.531914893617021,
"grad_norm": 1.83413124159608,
"learning_rate": 1.4513697938845571e-06,
"loss": 0.0329,
"step": 307
},
{
"epoch": 6.553191489361702,
"grad_norm": 1.9300831447222762,
"learning_rate": 1.4355517710873184e-06,
"loss": 0.0342,
"step": 308
},
{
"epoch": 6.574468085106383,
"grad_norm": 2.46498295305994,
"learning_rate": 1.419785627995044e-06,
"loss": 0.0293,
"step": 309
},
{
"epoch": 6.595744680851064,
"grad_norm": 1.8810365903428496,
"learning_rate": 1.4040721330273063e-06,
"loss": 0.0385,
"step": 310
},
{
"epoch": 6.617021276595745,
"grad_norm": 1.7341556708700896,
"learning_rate": 1.388412052037682e-06,
"loss": 0.0373,
"step": 311
},
{
"epoch": 6.638297872340425,
"grad_norm": 1.4886715538593713,
"learning_rate": 1.3728061482764238e-06,
"loss": 0.0374,
"step": 312
},
{
"epoch": 6.659574468085106,
"grad_norm": 1.8999495842705894,
"learning_rate": 1.3572551823532654e-06,
"loss": 0.0316,
"step": 313
},
{
"epoch": 6.680851063829787,
"grad_norm": 1.464463066404411,
"learning_rate": 1.3417599122003464e-06,
"loss": 0.033,
"step": 314
},
{
"epoch": 6.702127659574468,
"grad_norm": 1.8396892711483188,
"learning_rate": 1.3263210930352737e-06,
"loss": 0.0351,
"step": 315
},
{
"epoch": 6.723404255319149,
"grad_norm": 1.555502571877375,
"learning_rate": 1.3109394773243117e-06,
"loss": 0.0333,
"step": 316
},
{
"epoch": 6.74468085106383,
"grad_norm": 2.124808645505887,
"learning_rate": 1.2956158147457116e-06,
"loss": 0.0294,
"step": 317
},
{
"epoch": 6.76595744680851,
"grad_norm": 1.7688895712588661,
"learning_rate": 1.280350852153168e-06,
"loss": 0.0242,
"step": 318
},
{
"epoch": 6.787234042553192,
"grad_norm": 2.0383431308499325,
"learning_rate": 1.2651453335394232e-06,
"loss": 0.0353,
"step": 319
},
{
"epoch": 6.808510638297872,
"grad_norm": 1.7645071618559023,
"learning_rate": 1.2500000000000007e-06,
"loss": 0.0299,
"step": 320
},
{
"epoch": 6.829787234042553,
"grad_norm": 2.0474065053014217,
"learning_rate": 1.234915589697091e-06,
"loss": 0.0297,
"step": 321
},
{
"epoch": 6.851063829787234,
"grad_norm": 1.7823278045268314,
"learning_rate": 1.2198928378235717e-06,
"loss": 0.038,
"step": 322
},
{
"epoch": 6.872340425531915,
"grad_norm": 1.2380075941515003,
"learning_rate": 1.204932476567175e-06,
"loss": 0.031,
"step": 323
},
{
"epoch": 6.8936170212765955,
"grad_norm": 1.9796302395422996,
"learning_rate": 1.1900352350748026e-06,
"loss": 0.0381,
"step": 324
},
{
"epoch": 6.914893617021277,
"grad_norm": 1.3804831755312008,
"learning_rate": 1.1752018394169882e-06,
"loss": 0.0316,
"step": 325
},
{
"epoch": 6.9361702127659575,
"grad_norm": 1.6152383977065423,
"learning_rate": 1.160433012552508e-06,
"loss": 0.0343,
"step": 326
},
{
"epoch": 6.957446808510638,
"grad_norm": 1.69638635732096,
"learning_rate": 1.1457294742931508e-06,
"loss": 0.0296,
"step": 327
},
{
"epoch": 6.9787234042553195,
"grad_norm": 2.0806576931686456,
"learning_rate": 1.1310919412686248e-06,
"loss": 0.0341,
"step": 328
},
{
"epoch": 7.0,
"grad_norm": 1.8303598912826444,
"learning_rate": 1.11652112689164e-06,
"loss": 0.0405,
"step": 329
},
{
"epoch": 7.0212765957446805,
"grad_norm": 1.5859756930815514,
"learning_rate": 1.1020177413231334e-06,
"loss": 0.0252,
"step": 330
},
{
"epoch": 7.042553191489362,
"grad_norm": 0.7609936522020677,
"learning_rate": 1.0875824914376555e-06,
"loss": 0.0158,
"step": 331
},
{
"epoch": 7.0638297872340425,
"grad_norm": 2.0790528777967188,
"learning_rate": 1.073216080788921e-06,
"loss": 0.0175,
"step": 332
},
{
"epoch": 7.085106382978723,
"grad_norm": 1.0895054515077724,
"learning_rate": 1.0589192095755172e-06,
"loss": 0.0217,
"step": 333
},
{
"epoch": 7.1063829787234045,
"grad_norm": 1.0138202696131422,
"learning_rate": 1.0446925746067768e-06,
"loss": 0.0189,
"step": 334
},
{
"epoch": 7.127659574468085,
"grad_norm": 0.931336055385696,
"learning_rate": 1.0305368692688175e-06,
"loss": 0.0209,
"step": 335
},
{
"epoch": 7.148936170212766,
"grad_norm": 1.214374992403074,
"learning_rate": 1.0164527834907468e-06,
"loss": 0.0218,
"step": 336
},
{
"epoch": 7.170212765957447,
"grad_norm": 1.5245924187989146,
"learning_rate": 1.0024410037110358e-06,
"loss": 0.022,
"step": 337
},
{
"epoch": 7.191489361702128,
"grad_norm": 1.0820503643925237,
"learning_rate": 9.88502212844063e-07,
"loss": 0.0171,
"step": 338
},
{
"epoch": 7.212765957446808,
"grad_norm": 0.8009422999145568,
"learning_rate": 9.746370902468311e-07,
"loss": 0.0179,
"step": 339
},
{
"epoch": 7.23404255319149,
"grad_norm": 1.245836157318975,
"learning_rate": 9.608463116858544e-07,
"loss": 0.0196,
"step": 340
},
{
"epoch": 7.25531914893617,
"grad_norm": 0.9189114786673717,
"learning_rate": 9.471305493042243e-07,
"loss": 0.0206,
"step": 341
},
{
"epoch": 7.276595744680851,
"grad_norm": 1.2781854766390135,
"learning_rate": 9.334904715888496e-07,
"loss": 0.0164,
"step": 342
},
{
"epoch": 7.297872340425532,
"grad_norm": 1.5271374445001962,
"learning_rate": 9.199267433378728e-07,
"loss": 0.02,
"step": 343
},
{
"epoch": 7.319148936170213,
"grad_norm": 0.946456072976271,
"learning_rate": 9.064400256282757e-07,
"loss": 0.0214,
"step": 344
},
{
"epoch": 7.340425531914893,
"grad_norm": 1.270831674714735,
"learning_rate": 8.930309757836517e-07,
"loss": 0.0216,
"step": 345
},
{
"epoch": 7.361702127659575,
"grad_norm": 0.9172465899189646,
"learning_rate": 8.797002473421729e-07,
"loss": 0.0172,
"step": 346
},
{
"epoch": 7.382978723404255,
"grad_norm": 0.9762896610645739,
"learning_rate": 8.664484900247363e-07,
"loss": 0.0196,
"step": 347
},
{
"epoch": 7.404255319148936,
"grad_norm": 1.2366903890873284,
"learning_rate": 8.532763497032987e-07,
"loss": 0.0192,
"step": 348
},
{
"epoch": 7.425531914893617,
"grad_norm": 0.8921247183956026,
"learning_rate": 8.40184468369396e-07,
"loss": 0.0138,
"step": 349
},
{
"epoch": 7.446808510638298,
"grad_norm": 1.371683786414812,
"learning_rate": 8.271734841028553e-07,
"loss": 0.0257,
"step": 350
},
{
"epoch": 7.468085106382979,
"grad_norm": 1.6706279670454172,
"learning_rate": 8.142440310406923e-07,
"loss": 0.0221,
"step": 351
},
{
"epoch": 7.48936170212766,
"grad_norm": 1.0048836984452958,
"learning_rate": 8.013967393462094e-07,
"loss": 0.0179,
"step": 352
},
{
"epoch": 7.51063829787234,
"grad_norm": 1.028897861820127,
"learning_rate": 7.886322351782782e-07,
"loss": 0.0173,
"step": 353
},
{
"epoch": 7.531914893617021,
"grad_norm": 0.9414634888570663,
"learning_rate": 7.759511406608255e-07,
"loss": 0.0194,
"step": 354
},
{
"epoch": 7.553191489361702,
"grad_norm": 1.3421193366601667,
"learning_rate": 7.633540738525066e-07,
"loss": 0.0155,
"step": 355
},
{
"epoch": 7.574468085106383,
"grad_norm": 1.3991724028833572,
"learning_rate": 7.508416487165862e-07,
"loss": 0.0242,
"step": 356
},
{
"epoch": 7.595744680851064,
"grad_norm": 3.4908651855373067,
"learning_rate": 7.384144750910133e-07,
"loss": 0.0201,
"step": 357
},
{
"epoch": 7.617021276595745,
"grad_norm": 1.1125708607058544,
"learning_rate": 7.260731586586983e-07,
"loss": 0.0181,
"step": 358
},
{
"epoch": 7.638297872340425,
"grad_norm": 2.003422428134198,
"learning_rate": 7.138183009179922e-07,
"loss": 0.0244,
"step": 359
},
{
"epoch": 7.659574468085106,
"grad_norm": 0.8666069576765489,
"learning_rate": 7.016504991533727e-07,
"loss": 0.0231,
"step": 360
},
{
"epoch": 7.680851063829787,
"grad_norm": 0.8225857804515108,
"learning_rate": 6.895703464063319e-07,
"loss": 0.0127,
"step": 361
},
{
"epoch": 7.702127659574468,
"grad_norm": 1.2220101527746525,
"learning_rate": 6.775784314464717e-07,
"loss": 0.0261,
"step": 362
},
{
"epoch": 7.723404255319149,
"grad_norm": 1.291947063153676,
"learning_rate": 6.656753387428089e-07,
"loss": 0.024,
"step": 363
},
{
"epoch": 7.74468085106383,
"grad_norm": 1.288638995380391,
"learning_rate": 6.538616484352902e-07,
"loss": 0.016,
"step": 364
},
{
"epoch": 7.76595744680851,
"grad_norm": 1.3674238710009354,
"learning_rate": 6.421379363065142e-07,
"loss": 0.0253,
"step": 365
},
{
"epoch": 7.787234042553192,
"grad_norm": 0.877515831783811,
"learning_rate": 6.305047737536707e-07,
"loss": 0.0162,
"step": 366
},
{
"epoch": 7.808510638297872,
"grad_norm": 1.1630712659406794,
"learning_rate": 6.189627277606894e-07,
"loss": 0.0227,
"step": 367
},
{
"epoch": 7.829787234042553,
"grad_norm": 1.2379019290179472,
"learning_rate": 6.075123608706093e-07,
"loss": 0.0179,
"step": 368
},
{
"epoch": 7.851063829787234,
"grad_norm": 0.9761836229980341,
"learning_rate": 5.961542311581586e-07,
"loss": 0.017,
"step": 369
},
{
"epoch": 7.872340425531915,
"grad_norm": 1.2905446310752584,
"learning_rate": 5.848888922025553e-07,
"loss": 0.0237,
"step": 370
},
{
"epoch": 7.8936170212765955,
"grad_norm": 1.0599111161417107,
"learning_rate": 5.737168930605272e-07,
"loss": 0.0193,
"step": 371
},
{
"epoch": 7.914893617021277,
"grad_norm": 0.9002316779352258,
"learning_rate": 5.626387782395512e-07,
"loss": 0.0219,
"step": 372
},
{
"epoch": 7.9361702127659575,
"grad_norm": 1.0871181403304409,
"learning_rate": 5.516550876713142e-07,
"loss": 0.0202,
"step": 373
},
{
"epoch": 7.957446808510638,
"grad_norm": 0.7739867037107642,
"learning_rate": 5.407663566854008e-07,
"loss": 0.0141,
"step": 374
},
{
"epoch": 7.9787234042553195,
"grad_norm": 1.1646231525599027,
"learning_rate": 5.299731159831953e-07,
"loss": 0.0199,
"step": 375
},
{
"epoch": 8.0,
"grad_norm": 0.8486174354933951,
"learning_rate": 5.192758916120236e-07,
"loss": 0.0183,
"step": 376
},
{
"epoch": 8.02127659574468,
"grad_norm": 0.5857908765113855,
"learning_rate": 5.086752049395094e-07,
"loss": 0.0154,
"step": 377
},
{
"epoch": 8.042553191489361,
"grad_norm": 0.820073038781464,
"learning_rate": 4.981715726281666e-07,
"loss": 0.0162,
"step": 378
},
{
"epoch": 8.063829787234043,
"grad_norm": 0.5353750025209411,
"learning_rate": 4.87765506610215e-07,
"loss": 0.0117,
"step": 379
},
{
"epoch": 8.085106382978724,
"grad_norm": 0.6641723492880932,
"learning_rate": 4.774575140626317e-07,
"loss": 0.0155,
"step": 380
},
{
"epoch": 8.106382978723405,
"grad_norm": 1.5478055693073858,
"learning_rate": 4.672480973824312e-07,
"loss": 0.0138,
"step": 381
},
{
"epoch": 8.127659574468085,
"grad_norm": 0.5935452327954517,
"learning_rate": 4.5713775416217884e-07,
"loss": 0.0128,
"step": 382
},
{
"epoch": 8.148936170212766,
"grad_norm": 0.45835158236977624,
"learning_rate": 4.4712697716573994e-07,
"loss": 0.014,
"step": 383
},
{
"epoch": 8.170212765957446,
"grad_norm": 0.495302571001091,
"learning_rate": 4.372162543042624e-07,
"loss": 0.0135,
"step": 384
},
{
"epoch": 8.191489361702128,
"grad_norm": 0.4653447676593763,
"learning_rate": 4.27406068612396e-07,
"loss": 0.0113,
"step": 385
},
{
"epoch": 8.212765957446809,
"grad_norm": 0.5764443378886089,
"learning_rate": 4.1769689822475147e-07,
"loss": 0.0149,
"step": 386
},
{
"epoch": 8.23404255319149,
"grad_norm": 0.34507606047404416,
"learning_rate": 4.0808921635259595e-07,
"loss": 0.0096,
"step": 387
},
{
"epoch": 8.25531914893617,
"grad_norm": 0.6316918421422608,
"learning_rate": 3.9858349126078945e-07,
"loss": 0.0149,
"step": 388
},
{
"epoch": 8.27659574468085,
"grad_norm": 1.13945433310328,
"learning_rate": 3.891801862449629e-07,
"loss": 0.0171,
"step": 389
},
{
"epoch": 8.297872340425531,
"grad_norm": 1.6671178791582812,
"learning_rate": 3.798797596089351e-07,
"loss": 0.0162,
"step": 390
},
{
"epoch": 8.319148936170214,
"grad_norm": 0.6242884167930942,
"learning_rate": 3.7068266464238085e-07,
"loss": 0.0171,
"step": 391
},
{
"epoch": 8.340425531914894,
"grad_norm": 0.6157189166745389,
"learning_rate": 3.615893495987335e-07,
"loss": 0.0083,
"step": 392
},
{
"epoch": 8.361702127659575,
"grad_norm": 1.187375482456905,
"learning_rate": 3.5260025767333894e-07,
"loss": 0.0133,
"step": 393
},
{
"epoch": 8.382978723404255,
"grad_norm": 0.9198827823698973,
"learning_rate": 3.4371582698185636e-07,
"loss": 0.0168,
"step": 394
},
{
"epoch": 8.404255319148936,
"grad_norm": 0.9127337383828107,
"learning_rate": 3.3493649053890325e-07,
"loss": 0.0167,
"step": 395
},
{
"epoch": 8.425531914893616,
"grad_norm": 0.4716697636311403,
"learning_rate": 3.262626762369525e-07,
"loss": 0.0134,
"step": 396
},
{
"epoch": 8.446808510638299,
"grad_norm": 0.5430698963916597,
"learning_rate": 3.176948068254762e-07,
"loss": 0.0115,
"step": 397
},
{
"epoch": 8.46808510638298,
"grad_norm": 0.5589379122310724,
"learning_rate": 3.092332998903416e-07,
"loss": 0.0134,
"step": 398
},
{
"epoch": 8.48936170212766,
"grad_norm": 1.0297905449956246,
"learning_rate": 3.0087856783345916e-07,
"loss": 0.0165,
"step": 399
},
{
"epoch": 8.51063829787234,
"grad_norm": 0.5586126342363132,
"learning_rate": 2.9263101785268253e-07,
"loss": 0.0129,
"step": 400
},
{
"epoch": 8.51063829787234,
"eval_loss": 1.6673990488052368,
"eval_runtime": 1.6235,
"eval_samples_per_second": 14.167,
"eval_steps_per_second": 0.616,
"step": 400
},
{
"epoch": 8.53191489361702,
"grad_norm": 0.5428039786664497,
"learning_rate": 2.844910519219632e-07,
"loss": 0.0128,
"step": 401
},
{
"epoch": 8.553191489361701,
"grad_norm": 0.633307262716061,
"learning_rate": 2.764590667717562e-07,
"loss": 0.0161,
"step": 402
},
{
"epoch": 8.574468085106384,
"grad_norm": 0.5251963830034382,
"learning_rate": 2.6853545386968607e-07,
"loss": 0.0115,
"step": 403
},
{
"epoch": 8.595744680851064,
"grad_norm": 0.48124935494240734,
"learning_rate": 2.6072059940146775e-07,
"loss": 0.0097,
"step": 404
},
{
"epoch": 8.617021276595745,
"grad_norm": 0.5546850016440125,
"learning_rate": 2.53014884252083e-07,
"loss": 0.0095,
"step": 405
},
{
"epoch": 8.638297872340425,
"grad_norm": 0.6870597575288601,
"learning_rate": 2.454186839872158e-07,
"loss": 0.0161,
"step": 406
},
{
"epoch": 8.659574468085106,
"grad_norm": 0.5075138452670342,
"learning_rate": 2.3793236883495164e-07,
"loss": 0.0137,
"step": 407
},
{
"epoch": 8.680851063829786,
"grad_norm": 0.7184250889463067,
"learning_rate": 2.3055630366772857e-07,
"loss": 0.0124,
"step": 408
},
{
"epoch": 8.702127659574469,
"grad_norm": 0.6095068862526991,
"learning_rate": 2.2329084798455747e-07,
"loss": 0.0145,
"step": 409
},
{
"epoch": 8.72340425531915,
"grad_norm": 0.5513423746444825,
"learning_rate": 2.1613635589349756e-07,
"loss": 0.0155,
"step": 410
},
{
"epoch": 8.74468085106383,
"grad_norm": 0.5875898983901224,
"learning_rate": 2.0909317609440093e-07,
"loss": 0.0146,
"step": 411
},
{
"epoch": 8.76595744680851,
"grad_norm": 0.5750560265603247,
"learning_rate": 2.0216165186191406e-07,
"loss": 0.0131,
"step": 412
},
{
"epoch": 8.787234042553191,
"grad_norm": 0.6746724848758707,
"learning_rate": 1.95342121028749e-07,
"loss": 0.0118,
"step": 413
},
{
"epoch": 8.808510638297872,
"grad_norm": 0.6972301523231518,
"learning_rate": 1.8863491596921745e-07,
"loss": 0.0143,
"step": 414
},
{
"epoch": 8.829787234042554,
"grad_norm": 0.7525328974002997,
"learning_rate": 1.8204036358303173e-07,
"loss": 0.0144,
"step": 415
},
{
"epoch": 8.851063829787234,
"grad_norm": 0.5087036176440635,
"learning_rate": 1.7555878527937164e-07,
"loss": 0.0117,
"step": 416
},
{
"epoch": 8.872340425531915,
"grad_norm": 0.6061494141417032,
"learning_rate": 1.6919049696121957e-07,
"loss": 0.0168,
"step": 417
},
{
"epoch": 8.893617021276595,
"grad_norm": 0.6674633587572225,
"learning_rate": 1.629358090099639e-07,
"loss": 0.0154,
"step": 418
},
{
"epoch": 8.914893617021276,
"grad_norm": 0.6381527241007888,
"learning_rate": 1.567950262702714e-07,
"loss": 0.0118,
"step": 419
},
{
"epoch": 8.936170212765958,
"grad_norm": 0.5652826587480684,
"learning_rate": 1.507684480352292e-07,
"loss": 0.0142,
"step": 420
},
{
"epoch": 8.957446808510639,
"grad_norm": 0.5390136780396995,
"learning_rate": 1.4485636803175828e-07,
"loss": 0.0153,
"step": 421
},
{
"epoch": 8.97872340425532,
"grad_norm": 0.4433110433252025,
"learning_rate": 1.3905907440629752e-07,
"loss": 0.0128,
"step": 422
},
{
"epoch": 9.0,
"grad_norm": 0.3937254962258607,
"learning_rate": 1.3337684971075932e-07,
"loss": 0.0141,
"step": 423
},
{
"epoch": 9.02127659574468,
"grad_norm": 0.42872899914768886,
"learning_rate": 1.278099708887587e-07,
"loss": 0.0139,
"step": 424
},
{
"epoch": 9.042553191489361,
"grad_norm": 0.43176685764328526,
"learning_rate": 1.223587092621162e-07,
"loss": 0.01,
"step": 425
},
{
"epoch": 9.063829787234043,
"grad_norm": 0.3743485104670554,
"learning_rate": 1.1702333051763271e-07,
"loss": 0.0082,
"step": 426
},
{
"epoch": 9.085106382978724,
"grad_norm": 0.39987032601374245,
"learning_rate": 1.1180409469414094e-07,
"loss": 0.0125,
"step": 427
},
{
"epoch": 9.106382978723405,
"grad_norm": 0.39568260783662074,
"learning_rate": 1.067012561698319e-07,
"loss": 0.0099,
"step": 428
},
{
"epoch": 9.127659574468085,
"grad_norm": 0.4002164747878374,
"learning_rate": 1.0171506364985622e-07,
"loss": 0.0141,
"step": 429
},
{
"epoch": 9.148936170212766,
"grad_norm": 0.3140844748261339,
"learning_rate": 9.684576015420277e-08,
"loss": 0.0083,
"step": 430
},
{
"epoch": 9.170212765957446,
"grad_norm": 3.0839571989766026,
"learning_rate": 9.209358300585474e-08,
"loss": 0.0107,
"step": 431
},
{
"epoch": 9.191489361702128,
"grad_norm": 0.43386426711502496,
"learning_rate": 8.745876381922147e-08,
"loss": 0.0114,
"step": 432
},
{
"epoch": 9.212765957446809,
"grad_norm": 0.49575077577441945,
"learning_rate": 8.294152848885156e-08,
"loss": 0.0122,
"step": 433
},
{
"epoch": 9.23404255319149,
"grad_norm": 0.42605421543375827,
"learning_rate": 7.854209717842231e-08,
"loss": 0.0123,
"step": 434
},
{
"epoch": 9.25531914893617,
"grad_norm": 0.4159085198061852,
"learning_rate": 7.426068431000883e-08,
"loss": 0.011,
"step": 435
},
{
"epoch": 9.27659574468085,
"grad_norm": 0.4297510739797917,
"learning_rate": 7.009749855363457e-08,
"loss": 0.0096,
"step": 436
},
{
"epoch": 9.297872340425531,
"grad_norm": 0.39770284394181865,
"learning_rate": 6.605274281709929e-08,
"loss": 0.0108,
"step": 437
},
{
"epoch": 9.319148936170214,
"grad_norm": 0.3799358013058963,
"learning_rate": 6.212661423609184e-08,
"loss": 0.0135,
"step": 438
},
{
"epoch": 9.340425531914894,
"grad_norm": 0.40685286788703806,
"learning_rate": 5.83193041645802e-08,
"loss": 0.0105,
"step": 439
},
{
"epoch": 9.361702127659575,
"grad_norm": 0.5276489442708509,
"learning_rate": 5.463099816548578e-08,
"loss": 0.0159,
"step": 440
},
{
"epoch": 9.382978723404255,
"grad_norm": 0.3398816914667588,
"learning_rate": 5.106187600163987e-08,
"loss": 0.01,
"step": 441
},
{
"epoch": 9.404255319148936,
"grad_norm": 0.43060715328426347,
"learning_rate": 4.761211162702117e-08,
"loss": 0.0109,
"step": 442
},
{
"epoch": 9.425531914893616,
"grad_norm": 0.6155281563788073,
"learning_rate": 4.428187317827848e-08,
"loss": 0.0107,
"step": 443
},
{
"epoch": 9.446808510638299,
"grad_norm": 0.4920360355401412,
"learning_rate": 4.1071322966535487e-08,
"loss": 0.0139,
"step": 444
},
{
"epoch": 9.46808510638298,
"grad_norm": 0.44825099453614115,
"learning_rate": 3.798061746947995e-08,
"loss": 0.0113,
"step": 445
},
{
"epoch": 9.48936170212766,
"grad_norm": 0.48315881463136157,
"learning_rate": 3.5009907323737826e-08,
"loss": 0.0131,
"step": 446
},
{
"epoch": 9.51063829787234,
"grad_norm": 0.3897061025707457,
"learning_rate": 3.2159337317530234e-08,
"loss": 0.0131,
"step": 447
},
{
"epoch": 9.53191489361702,
"grad_norm": 0.3845441440160021,
"learning_rate": 2.9429046383618042e-08,
"loss": 0.0094,
"step": 448
},
{
"epoch": 9.553191489361701,
"grad_norm": 0.4018776472448016,
"learning_rate": 2.681916759252917e-08,
"loss": 0.0117,
"step": 449
},
{
"epoch": 9.574468085106384,
"grad_norm": 0.3789922116838833,
"learning_rate": 2.4329828146074096e-08,
"loss": 0.0099,
"step": 450
},
{
"epoch": 9.595744680851064,
"grad_norm": 0.47356186931300714,
"learning_rate": 2.1961149371145795e-08,
"loss": 0.012,
"step": 451
},
{
"epoch": 9.617021276595745,
"grad_norm": 0.5175691416116797,
"learning_rate": 1.9713246713805588e-08,
"loss": 0.0137,
"step": 452
},
{
"epoch": 9.638297872340425,
"grad_norm": 0.35550871415754165,
"learning_rate": 1.7586229733657646e-08,
"loss": 0.0126,
"step": 453
},
{
"epoch": 9.659574468085106,
"grad_norm": 0.4521940543192018,
"learning_rate": 1.5580202098509078e-08,
"loss": 0.0123,
"step": 454
},
{
"epoch": 9.680851063829786,
"grad_norm": 0.3645510186204655,
"learning_rate": 1.3695261579316776e-08,
"loss": 0.0099,
"step": 455
},
{
"epoch": 9.702127659574469,
"grad_norm": 0.5052016943572653,
"learning_rate": 1.193150004542204e-08,
"loss": 0.0108,
"step": 456
},
{
"epoch": 9.72340425531915,
"grad_norm": 0.4591420783076054,
"learning_rate": 1.0289003460074165e-08,
"loss": 0.0136,
"step": 457
},
{
"epoch": 9.74468085106383,
"grad_norm": 0.3979451604394145,
"learning_rate": 8.767851876239075e-09,
"loss": 0.0124,
"step": 458
},
{
"epoch": 9.76595744680851,
"grad_norm": 0.4508311448933257,
"learning_rate": 7.368119432699383e-09,
"loss": 0.0117,
"step": 459
},
{
"epoch": 9.787234042553191,
"grad_norm": 0.48162461236003445,
"learning_rate": 6.089874350439507e-09,
"loss": 0.0131,
"step": 460
},
{
"epoch": 9.808510638297872,
"grad_norm": 0.3727617520322409,
"learning_rate": 4.933178929321103e-09,
"loss": 0.0092,
"step": 461
},
{
"epoch": 9.829787234042554,
"grad_norm": 0.37748204227545856,
"learning_rate": 3.8980895450474455e-09,
"loss": 0.0108,
"step": 462
},
{
"epoch": 9.851063829787234,
"grad_norm": 0.30954369790579545,
"learning_rate": 2.984656646415063e-09,
"loss": 0.0082,
"step": 463
},
{
"epoch": 9.872340425531915,
"grad_norm": 0.441565820890093,
"learning_rate": 2.192924752854042e-09,
"loss": 0.014,
"step": 464
},
{
"epoch": 9.893617021276595,
"grad_norm": 0.42642106144925446,
"learning_rate": 1.5229324522605949e-09,
"loss": 0.0086,
"step": 465
},
{
"epoch": 9.914893617021276,
"grad_norm": 0.3393626098420996,
"learning_rate": 9.747123991141193e-10,
"loss": 0.008,
"step": 466
},
{
"epoch": 9.936170212765958,
"grad_norm": 0.4441469010998739,
"learning_rate": 5.48291312886251e-10,
"loss": 0.0135,
"step": 467
},
{
"epoch": 9.957446808510639,
"grad_norm": 0.44011051414564956,
"learning_rate": 2.43689976739403e-10,
"loss": 0.0117,
"step": 468
},
{
"epoch": 9.97872340425532,
"grad_norm": 0.47111655728750834,
"learning_rate": 6.092323651313293e-11,
"loss": 0.0104,
"step": 469
},
{
"epoch": 10.0,
"grad_norm": 0.5133248194623898,
"learning_rate": 0.0,
"loss": 0.0161,
"step": 470
}
],
"logging_steps": 1.0,
"max_steps": 470,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 32811979898880.0,
"train_batch_size": 6,
"trial_name": null,
"trial_params": null
}