3340 lines
80 KiB
JSON
3340 lines
80 KiB
JSON
{
|
|
"best_metric": null,
|
|
"best_model_checkpoint": null,
|
|
"epoch": 10.0,
|
|
"eval_steps": 200,
|
|
"global_step": 470,
|
|
"is_hyper_param_search": false,
|
|
"is_local_process_zero": true,
|
|
"is_world_process_zero": true,
|
|
"log_history": [
|
|
{
|
|
"epoch": 0.02127659574468085,
|
|
"grad_norm": 19.769559521884688,
|
|
"learning_rate": 2.5000000000000004e-07,
|
|
"loss": 1.5136,
|
|
"step": 1
|
|
},
|
|
{
|
|
"epoch": 0.0425531914893617,
|
|
"grad_norm": 22.366909847966127,
|
|
"learning_rate": 5.000000000000001e-07,
|
|
"loss": 1.4656,
|
|
"step": 2
|
|
},
|
|
{
|
|
"epoch": 0.06382978723404255,
|
|
"grad_norm": 27.397989604017898,
|
|
"learning_rate": 7.5e-07,
|
|
"loss": 1.5124,
|
|
"step": 3
|
|
},
|
|
{
|
|
"epoch": 0.0851063829787234,
|
|
"grad_norm": 18.58567583552412,
|
|
"learning_rate": 1.0000000000000002e-06,
|
|
"loss": 1.5225,
|
|
"step": 4
|
|
},
|
|
{
|
|
"epoch": 0.10638297872340426,
|
|
"grad_norm": 18.165996038203307,
|
|
"learning_rate": 1.25e-06,
|
|
"loss": 1.5576,
|
|
"step": 5
|
|
},
|
|
{
|
|
"epoch": 0.1276595744680851,
|
|
"grad_norm": 13.681392535187342,
|
|
"learning_rate": 1.5e-06,
|
|
"loss": 1.4687,
|
|
"step": 6
|
|
},
|
|
{
|
|
"epoch": 0.14893617021276595,
|
|
"grad_norm": 14.999691223127021,
|
|
"learning_rate": 1.75e-06,
|
|
"loss": 1.3933,
|
|
"step": 7
|
|
},
|
|
{
|
|
"epoch": 0.1702127659574468,
|
|
"grad_norm": 9.360244090412072,
|
|
"learning_rate": 2.0000000000000003e-06,
|
|
"loss": 1.4721,
|
|
"step": 8
|
|
},
|
|
{
|
|
"epoch": 0.19148936170212766,
|
|
"grad_norm": 11.174023539278728,
|
|
"learning_rate": 2.25e-06,
|
|
"loss": 1.44,
|
|
"step": 9
|
|
},
|
|
{
|
|
"epoch": 0.2127659574468085,
|
|
"grad_norm": 9.866648780480253,
|
|
"learning_rate": 2.5e-06,
|
|
"loss": 1.2595,
|
|
"step": 10
|
|
},
|
|
{
|
|
"epoch": 0.23404255319148937,
|
|
"grad_norm": 9.512446051741868,
|
|
"learning_rate": 2.7500000000000004e-06,
|
|
"loss": 1.4442,
|
|
"step": 11
|
|
},
|
|
{
|
|
"epoch": 0.2553191489361702,
|
|
"grad_norm": 8.670274331402473,
|
|
"learning_rate": 3e-06,
|
|
"loss": 1.2739,
|
|
"step": 12
|
|
},
|
|
{
|
|
"epoch": 0.2765957446808511,
|
|
"grad_norm": 6.920817280642445,
|
|
"learning_rate": 3.2500000000000002e-06,
|
|
"loss": 1.276,
|
|
"step": 13
|
|
},
|
|
{
|
|
"epoch": 0.2978723404255319,
|
|
"grad_norm": 7.10414215318037,
|
|
"learning_rate": 3.5e-06,
|
|
"loss": 1.2421,
|
|
"step": 14
|
|
},
|
|
{
|
|
"epoch": 0.3191489361702128,
|
|
"grad_norm": 6.159725671936445,
|
|
"learning_rate": 3.7500000000000005e-06,
|
|
"loss": 1.239,
|
|
"step": 15
|
|
},
|
|
{
|
|
"epoch": 0.3404255319148936,
|
|
"grad_norm": 7.6873175205954825,
|
|
"learning_rate": 4.000000000000001e-06,
|
|
"loss": 1.2396,
|
|
"step": 16
|
|
},
|
|
{
|
|
"epoch": 0.3617021276595745,
|
|
"grad_norm": 6.528518611068642,
|
|
"learning_rate": 4.25e-06,
|
|
"loss": 1.1543,
|
|
"step": 17
|
|
},
|
|
{
|
|
"epoch": 0.3829787234042553,
|
|
"grad_norm": 6.77428307591927,
|
|
"learning_rate": 4.5e-06,
|
|
"loss": 1.1545,
|
|
"step": 18
|
|
},
|
|
{
|
|
"epoch": 0.40425531914893614,
|
|
"grad_norm": 6.417164213417,
|
|
"learning_rate": 4.75e-06,
|
|
"loss": 1.0959,
|
|
"step": 19
|
|
},
|
|
{
|
|
"epoch": 0.425531914893617,
|
|
"grad_norm": 6.180266176299509,
|
|
"learning_rate": 5e-06,
|
|
"loss": 1.1381,
|
|
"step": 20
|
|
},
|
|
{
|
|
"epoch": 0.44680851063829785,
|
|
"grad_norm": 6.238735118043302,
|
|
"learning_rate": 4.999939076763487e-06,
|
|
"loss": 0.9938,
|
|
"step": 21
|
|
},
|
|
{
|
|
"epoch": 0.46808510638297873,
|
|
"grad_norm": 5.503542713382496,
|
|
"learning_rate": 4.999756310023261e-06,
|
|
"loss": 1.0336,
|
|
"step": 22
|
|
},
|
|
{
|
|
"epoch": 0.48936170212765956,
|
|
"grad_norm": 5.7109281559241305,
|
|
"learning_rate": 4.999451708687114e-06,
|
|
"loss": 1.0293,
|
|
"step": 23
|
|
},
|
|
{
|
|
"epoch": 0.5106382978723404,
|
|
"grad_norm": 5.933533265865748,
|
|
"learning_rate": 4.999025287600886e-06,
|
|
"loss": 1.0838,
|
|
"step": 24
|
|
},
|
|
{
|
|
"epoch": 0.5319148936170213,
|
|
"grad_norm": 6.058382477727568,
|
|
"learning_rate": 4.99847706754774e-06,
|
|
"loss": 1.147,
|
|
"step": 25
|
|
},
|
|
{
|
|
"epoch": 0.5531914893617021,
|
|
"grad_norm": 5.039016064603872,
|
|
"learning_rate": 4.997807075247147e-06,
|
|
"loss": 1.0661,
|
|
"step": 26
|
|
},
|
|
{
|
|
"epoch": 0.574468085106383,
|
|
"grad_norm": 5.673653311719194,
|
|
"learning_rate": 4.9970153433535855e-06,
|
|
"loss": 1.1088,
|
|
"step": 27
|
|
},
|
|
{
|
|
"epoch": 0.5957446808510638,
|
|
"grad_norm": 6.772514217020031,
|
|
"learning_rate": 4.996101910454953e-06,
|
|
"loss": 1.0398,
|
|
"step": 28
|
|
},
|
|
{
|
|
"epoch": 0.6170212765957447,
|
|
"grad_norm": 7.261784604470834,
|
|
"learning_rate": 4.9950668210706795e-06,
|
|
"loss": 1.1197,
|
|
"step": 29
|
|
},
|
|
{
|
|
"epoch": 0.6382978723404256,
|
|
"grad_norm": 5.125305508590899,
|
|
"learning_rate": 4.993910125649561e-06,
|
|
"loss": 1.0096,
|
|
"step": 30
|
|
},
|
|
{
|
|
"epoch": 0.6595744680851063,
|
|
"grad_norm": 5.79320276945061,
|
|
"learning_rate": 4.992631880567301e-06,
|
|
"loss": 1.0418,
|
|
"step": 31
|
|
},
|
|
{
|
|
"epoch": 0.6808510638297872,
|
|
"grad_norm": 6.583723424976928,
|
|
"learning_rate": 4.9912321481237616e-06,
|
|
"loss": 1.1084,
|
|
"step": 32
|
|
},
|
|
{
|
|
"epoch": 0.7021276595744681,
|
|
"grad_norm": 5.166238541818618,
|
|
"learning_rate": 4.989710996539926e-06,
|
|
"loss": 1.0269,
|
|
"step": 33
|
|
},
|
|
{
|
|
"epoch": 0.723404255319149,
|
|
"grad_norm": 4.702597984819574,
|
|
"learning_rate": 4.988068499954578e-06,
|
|
"loss": 0.9362,
|
|
"step": 34
|
|
},
|
|
{
|
|
"epoch": 0.7446808510638298,
|
|
"grad_norm": 5.035553587199529,
|
|
"learning_rate": 4.986304738420684e-06,
|
|
"loss": 1.0027,
|
|
"step": 35
|
|
},
|
|
{
|
|
"epoch": 0.7659574468085106,
|
|
"grad_norm": 5.979469659293992,
|
|
"learning_rate": 4.984419797901491e-06,
|
|
"loss": 1.0197,
|
|
"step": 36
|
|
},
|
|
{
|
|
"epoch": 0.7872340425531915,
|
|
"grad_norm": 7.251693613828905,
|
|
"learning_rate": 4.9824137702663424e-06,
|
|
"loss": 1.0665,
|
|
"step": 37
|
|
},
|
|
{
|
|
"epoch": 0.8085106382978723,
|
|
"grad_norm": 5.336704026070371,
|
|
"learning_rate": 4.980286753286196e-06,
|
|
"loss": 1.0612,
|
|
"step": 38
|
|
},
|
|
{
|
|
"epoch": 0.8297872340425532,
|
|
"grad_norm": 4.572621845298497,
|
|
"learning_rate": 4.978038850628855e-06,
|
|
"loss": 0.891,
|
|
"step": 39
|
|
},
|
|
{
|
|
"epoch": 0.851063829787234,
|
|
"grad_norm": 5.901358668805014,
|
|
"learning_rate": 4.975670171853926e-06,
|
|
"loss": 1.003,
|
|
"step": 40
|
|
},
|
|
{
|
|
"epoch": 0.8723404255319149,
|
|
"grad_norm": 5.967637194891353,
|
|
"learning_rate": 4.973180832407471e-06,
|
|
"loss": 1.0412,
|
|
"step": 41
|
|
},
|
|
{
|
|
"epoch": 0.8936170212765957,
|
|
"grad_norm": 4.651650771697547,
|
|
"learning_rate": 4.970570953616383e-06,
|
|
"loss": 0.905,
|
|
"step": 42
|
|
},
|
|
{
|
|
"epoch": 0.9148936170212766,
|
|
"grad_norm": 4.40595275321189,
|
|
"learning_rate": 4.96784066268247e-06,
|
|
"loss": 0.9606,
|
|
"step": 43
|
|
},
|
|
{
|
|
"epoch": 0.9361702127659575,
|
|
"grad_norm": 4.297391018229843,
|
|
"learning_rate": 4.964990092676263e-06,
|
|
"loss": 0.912,
|
|
"step": 44
|
|
},
|
|
{
|
|
"epoch": 0.9574468085106383,
|
|
"grad_norm": 5.8520797403685565,
|
|
"learning_rate": 4.962019382530521e-06,
|
|
"loss": 0.99,
|
|
"step": 45
|
|
},
|
|
{
|
|
"epoch": 0.9787234042553191,
|
|
"grad_norm": 4.460582020214977,
|
|
"learning_rate": 4.958928677033465e-06,
|
|
"loss": 0.9338,
|
|
"step": 46
|
|
},
|
|
{
|
|
"epoch": 1.0,
|
|
"grad_norm": 5.108150405797771,
|
|
"learning_rate": 4.9557181268217225e-06,
|
|
"loss": 0.8841,
|
|
"step": 47
|
|
},
|
|
{
|
|
"epoch": 1.0212765957446808,
|
|
"grad_norm": 4.42640489710439,
|
|
"learning_rate": 4.9523878883729794e-06,
|
|
"loss": 0.6441,
|
|
"step": 48
|
|
},
|
|
{
|
|
"epoch": 1.0425531914893618,
|
|
"grad_norm": 4.783433777129786,
|
|
"learning_rate": 4.94893812399836e-06,
|
|
"loss": 0.7076,
|
|
"step": 49
|
|
},
|
|
{
|
|
"epoch": 1.0638297872340425,
|
|
"grad_norm": 4.461478664530093,
|
|
"learning_rate": 4.9453690018345144e-06,
|
|
"loss": 0.6507,
|
|
"step": 50
|
|
},
|
|
{
|
|
"epoch": 1.0851063829787233,
|
|
"grad_norm": 4.659068920568944,
|
|
"learning_rate": 4.9416806958354206e-06,
|
|
"loss": 0.6388,
|
|
"step": 51
|
|
},
|
|
{
|
|
"epoch": 1.1063829787234043,
|
|
"grad_norm": 5.326272476452319,
|
|
"learning_rate": 4.937873385763909e-06,
|
|
"loss": 0.6629,
|
|
"step": 52
|
|
},
|
|
{
|
|
"epoch": 1.127659574468085,
|
|
"grad_norm": 4.42497697146127,
|
|
"learning_rate": 4.933947257182901e-06,
|
|
"loss": 0.6969,
|
|
"step": 53
|
|
},
|
|
{
|
|
"epoch": 1.148936170212766,
|
|
"grad_norm": 4.342156315066868,
|
|
"learning_rate": 4.9299025014463665e-06,
|
|
"loss": 0.6389,
|
|
"step": 54
|
|
},
|
|
{
|
|
"epoch": 1.1702127659574468,
|
|
"grad_norm": 5.535251688604961,
|
|
"learning_rate": 4.925739315689991e-06,
|
|
"loss": 0.6549,
|
|
"step": 55
|
|
},
|
|
{
|
|
"epoch": 1.1914893617021276,
|
|
"grad_norm": 4.098290419568537,
|
|
"learning_rate": 4.921457902821578e-06,
|
|
"loss": 0.6581,
|
|
"step": 56
|
|
},
|
|
{
|
|
"epoch": 1.2127659574468086,
|
|
"grad_norm": 4.804191552305714,
|
|
"learning_rate": 4.917058471511149e-06,
|
|
"loss": 0.6574,
|
|
"step": 57
|
|
},
|
|
{
|
|
"epoch": 1.2340425531914894,
|
|
"grad_norm": 4.661998643228261,
|
|
"learning_rate": 4.912541236180779e-06,
|
|
"loss": 0.6206,
|
|
"step": 58
|
|
},
|
|
{
|
|
"epoch": 1.2553191489361701,
|
|
"grad_norm": 5.457436437790183,
|
|
"learning_rate": 4.907906416994146e-06,
|
|
"loss": 0.6949,
|
|
"step": 59
|
|
},
|
|
{
|
|
"epoch": 1.2765957446808511,
|
|
"grad_norm": 5.287463616155875,
|
|
"learning_rate": 4.903154239845798e-06,
|
|
"loss": 0.6171,
|
|
"step": 60
|
|
},
|
|
{
|
|
"epoch": 1.297872340425532,
|
|
"grad_norm": 6.534162370166582,
|
|
"learning_rate": 4.898284936350144e-06,
|
|
"loss": 0.6738,
|
|
"step": 61
|
|
},
|
|
{
|
|
"epoch": 1.3191489361702127,
|
|
"grad_norm": 5.14242301334199,
|
|
"learning_rate": 4.893298743830168e-06,
|
|
"loss": 0.71,
|
|
"step": 62
|
|
},
|
|
{
|
|
"epoch": 1.3404255319148937,
|
|
"grad_norm": 4.864653117677477,
|
|
"learning_rate": 4.888195905305859e-06,
|
|
"loss": 0.6353,
|
|
"step": 63
|
|
},
|
|
{
|
|
"epoch": 1.3617021276595744,
|
|
"grad_norm": 4.469502230256778,
|
|
"learning_rate": 4.882976669482368e-06,
|
|
"loss": 0.6301,
|
|
"step": 64
|
|
},
|
|
{
|
|
"epoch": 1.3829787234042552,
|
|
"grad_norm": 4.650400360608429,
|
|
"learning_rate": 4.8776412907378845e-06,
|
|
"loss": 0.6298,
|
|
"step": 65
|
|
},
|
|
{
|
|
"epoch": 1.4042553191489362,
|
|
"grad_norm": 4.75068049245489,
|
|
"learning_rate": 4.8721900291112415e-06,
|
|
"loss": 0.6132,
|
|
"step": 66
|
|
},
|
|
{
|
|
"epoch": 1.425531914893617,
|
|
"grad_norm": 5.406681668582622,
|
|
"learning_rate": 4.866623150289241e-06,
|
|
"loss": 0.6177,
|
|
"step": 67
|
|
},
|
|
{
|
|
"epoch": 1.4468085106382977,
|
|
"grad_norm": 4.689427420330228,
|
|
"learning_rate": 4.860940925593703e-06,
|
|
"loss": 0.601,
|
|
"step": 68
|
|
},
|
|
{
|
|
"epoch": 1.4680851063829787,
|
|
"grad_norm": 5.66362626435671,
|
|
"learning_rate": 4.855143631968242e-06,
|
|
"loss": 0.6987,
|
|
"step": 69
|
|
},
|
|
{
|
|
"epoch": 1.4893617021276595,
|
|
"grad_norm": 6.639330863120106,
|
|
"learning_rate": 4.849231551964771e-06,
|
|
"loss": 0.705,
|
|
"step": 70
|
|
},
|
|
{
|
|
"epoch": 1.5106382978723403,
|
|
"grad_norm": 5.15079243645691,
|
|
"learning_rate": 4.84320497372973e-06,
|
|
"loss": 0.6812,
|
|
"step": 71
|
|
},
|
|
{
|
|
"epoch": 1.5319148936170213,
|
|
"grad_norm": 5.318900788335571,
|
|
"learning_rate": 4.837064190990036e-06,
|
|
"loss": 0.6717,
|
|
"step": 72
|
|
},
|
|
{
|
|
"epoch": 1.5531914893617023,
|
|
"grad_norm": 4.39050844628742,
|
|
"learning_rate": 4.830809503038781e-06,
|
|
"loss": 0.5978,
|
|
"step": 73
|
|
},
|
|
{
|
|
"epoch": 1.574468085106383,
|
|
"grad_norm": 4.11981361882099,
|
|
"learning_rate": 4.824441214720629e-06,
|
|
"loss": 0.5789,
|
|
"step": 74
|
|
},
|
|
{
|
|
"epoch": 1.5957446808510638,
|
|
"grad_norm": 4.898106888331966,
|
|
"learning_rate": 4.817959636416969e-06,
|
|
"loss": 0.638,
|
|
"step": 75
|
|
},
|
|
{
|
|
"epoch": 1.6170212765957448,
|
|
"grad_norm": 4.144574853637901,
|
|
"learning_rate": 4.811365084030784e-06,
|
|
"loss": 0.6631,
|
|
"step": 76
|
|
},
|
|
{
|
|
"epoch": 1.6382978723404256,
|
|
"grad_norm": 4.236596267103849,
|
|
"learning_rate": 4.804657878971252e-06,
|
|
"loss": 0.5618,
|
|
"step": 77
|
|
},
|
|
{
|
|
"epoch": 1.6595744680851063,
|
|
"grad_norm": 4.772661867276851,
|
|
"learning_rate": 4.7978383481380865e-06,
|
|
"loss": 0.6064,
|
|
"step": 78
|
|
},
|
|
{
|
|
"epoch": 1.6808510638297873,
|
|
"grad_norm": 5.5176007762843335,
|
|
"learning_rate": 4.790906823905599e-06,
|
|
"loss": 0.6346,
|
|
"step": 79
|
|
},
|
|
{
|
|
"epoch": 1.702127659574468,
|
|
"grad_norm": 4.1371187239001825,
|
|
"learning_rate": 4.783863644106502e-06,
|
|
"loss": 0.598,
|
|
"step": 80
|
|
},
|
|
{
|
|
"epoch": 1.7234042553191489,
|
|
"grad_norm": 4.617726148114806,
|
|
"learning_rate": 4.776709152015443e-06,
|
|
"loss": 0.572,
|
|
"step": 81
|
|
},
|
|
{
|
|
"epoch": 1.7446808510638299,
|
|
"grad_norm": 4.389883430040871,
|
|
"learning_rate": 4.769443696332272e-06,
|
|
"loss": 0.6145,
|
|
"step": 82
|
|
},
|
|
{
|
|
"epoch": 1.7659574468085106,
|
|
"grad_norm": 5.027199601611735,
|
|
"learning_rate": 4.762067631165049e-06,
|
|
"loss": 0.6858,
|
|
"step": 83
|
|
},
|
|
{
|
|
"epoch": 1.7872340425531914,
|
|
"grad_norm": 4.990601804039334,
|
|
"learning_rate": 4.754581316012785e-06,
|
|
"loss": 0.561,
|
|
"step": 84
|
|
},
|
|
{
|
|
"epoch": 1.8085106382978724,
|
|
"grad_norm": 5.705018779334882,
|
|
"learning_rate": 4.746985115747918e-06,
|
|
"loss": 0.6323,
|
|
"step": 85
|
|
},
|
|
{
|
|
"epoch": 1.8297872340425532,
|
|
"grad_norm": 5.1505009350987265,
|
|
"learning_rate": 4.7392794005985324e-06,
|
|
"loss": 0.5952,
|
|
"step": 86
|
|
},
|
|
{
|
|
"epoch": 1.851063829787234,
|
|
"grad_norm": 5.16517962200903,
|
|
"learning_rate": 4.731464546130315e-06,
|
|
"loss": 0.6462,
|
|
"step": 87
|
|
},
|
|
{
|
|
"epoch": 1.872340425531915,
|
|
"grad_norm": 5.203202926098486,
|
|
"learning_rate": 4.723540933228245e-06,
|
|
"loss": 0.5949,
|
|
"step": 88
|
|
},
|
|
{
|
|
"epoch": 1.8936170212765957,
|
|
"grad_norm": 4.6706368589871134,
|
|
"learning_rate": 4.715508948078037e-06,
|
|
"loss": 0.5878,
|
|
"step": 89
|
|
},
|
|
{
|
|
"epoch": 1.9148936170212765,
|
|
"grad_norm": 4.351714864184194,
|
|
"learning_rate": 4.707368982147318e-06,
|
|
"loss": 0.6227,
|
|
"step": 90
|
|
},
|
|
{
|
|
"epoch": 1.9361702127659575,
|
|
"grad_norm": 5.11397965814302,
|
|
"learning_rate": 4.699121432166542e-06,
|
|
"loss": 0.6347,
|
|
"step": 91
|
|
},
|
|
{
|
|
"epoch": 1.9574468085106385,
|
|
"grad_norm": 5.875205347984843,
|
|
"learning_rate": 4.690766700109659e-06,
|
|
"loss": 0.5905,
|
|
"step": 92
|
|
},
|
|
{
|
|
"epoch": 1.978723404255319,
|
|
"grad_norm": 6.03539957826197,
|
|
"learning_rate": 4.682305193174524e-06,
|
|
"loss": 0.6894,
|
|
"step": 93
|
|
},
|
|
{
|
|
"epoch": 2.0,
|
|
"grad_norm": 3.8707434069223026,
|
|
"learning_rate": 4.673737323763048e-06,
|
|
"loss": 0.4095,
|
|
"step": 94
|
|
},
|
|
{
|
|
"epoch": 2.021276595744681,
|
|
"grad_norm": 4.967672001558827,
|
|
"learning_rate": 4.665063509461098e-06,
|
|
"loss": 0.3403,
|
|
"step": 95
|
|
},
|
|
{
|
|
"epoch": 2.0425531914893615,
|
|
"grad_norm": 3.5072435391900436,
|
|
"learning_rate": 4.656284173018144e-06,
|
|
"loss": 0.3736,
|
|
"step": 96
|
|
},
|
|
{
|
|
"epoch": 2.0638297872340425,
|
|
"grad_norm": 4.360376932941338,
|
|
"learning_rate": 4.6473997423266615e-06,
|
|
"loss": 0.4061,
|
|
"step": 97
|
|
},
|
|
{
|
|
"epoch": 2.0851063829787235,
|
|
"grad_norm": 4.224245961638772,
|
|
"learning_rate": 4.638410650401267e-06,
|
|
"loss": 0.3515,
|
|
"step": 98
|
|
},
|
|
{
|
|
"epoch": 2.106382978723404,
|
|
"grad_norm": 4.792845268520252,
|
|
"learning_rate": 4.62931733535762e-06,
|
|
"loss": 0.3432,
|
|
"step": 99
|
|
},
|
|
{
|
|
"epoch": 2.127659574468085,
|
|
"grad_norm": 4.9719261969262085,
|
|
"learning_rate": 4.620120240391065e-06,
|
|
"loss": 0.3495,
|
|
"step": 100
|
|
},
|
|
{
|
|
"epoch": 2.148936170212766,
|
|
"grad_norm": 3.9393952177630864,
|
|
"learning_rate": 4.610819813755038e-06,
|
|
"loss": 0.3505,
|
|
"step": 101
|
|
},
|
|
{
|
|
"epoch": 2.1702127659574466,
|
|
"grad_norm": 5.0676129989734084,
|
|
"learning_rate": 4.601416508739211e-06,
|
|
"loss": 0.345,
|
|
"step": 102
|
|
},
|
|
{
|
|
"epoch": 2.1914893617021276,
|
|
"grad_norm": 4.268873683189594,
|
|
"learning_rate": 4.591910783647405e-06,
|
|
"loss": 0.3411,
|
|
"step": 103
|
|
},
|
|
{
|
|
"epoch": 2.2127659574468086,
|
|
"grad_norm": 5.648372477585861,
|
|
"learning_rate": 4.582303101775249e-06,
|
|
"loss": 0.3414,
|
|
"step": 104
|
|
},
|
|
{
|
|
"epoch": 2.2340425531914896,
|
|
"grad_norm": 4.569109738179888,
|
|
"learning_rate": 4.572593931387604e-06,
|
|
"loss": 0.3107,
|
|
"step": 105
|
|
},
|
|
{
|
|
"epoch": 2.25531914893617,
|
|
"grad_norm": 5.986604177134095,
|
|
"learning_rate": 4.562783745695738e-06,
|
|
"loss": 0.358,
|
|
"step": 106
|
|
},
|
|
{
|
|
"epoch": 2.276595744680851,
|
|
"grad_norm": 4.593476704693552,
|
|
"learning_rate": 4.55287302283426e-06,
|
|
"loss": 0.3938,
|
|
"step": 107
|
|
},
|
|
{
|
|
"epoch": 2.297872340425532,
|
|
"grad_norm": 4.729132719197455,
|
|
"learning_rate": 4.542862245837821e-06,
|
|
"loss": 0.3445,
|
|
"step": 108
|
|
},
|
|
{
|
|
"epoch": 2.3191489361702127,
|
|
"grad_norm": 5.319389221723198,
|
|
"learning_rate": 4.5327519026175694e-06,
|
|
"loss": 0.3638,
|
|
"step": 109
|
|
},
|
|
{
|
|
"epoch": 2.3404255319148937,
|
|
"grad_norm": 4.9962223266535295,
|
|
"learning_rate": 4.522542485937369e-06,
|
|
"loss": 0.3151,
|
|
"step": 110
|
|
},
|
|
{
|
|
"epoch": 2.3617021276595747,
|
|
"grad_norm": 5.43349545021367,
|
|
"learning_rate": 4.512234493389785e-06,
|
|
"loss": 0.3386,
|
|
"step": 111
|
|
},
|
|
{
|
|
"epoch": 2.382978723404255,
|
|
"grad_norm": 4.228899687733491,
|
|
"learning_rate": 4.501828427371834e-06,
|
|
"loss": 0.3381,
|
|
"step": 112
|
|
},
|
|
{
|
|
"epoch": 2.404255319148936,
|
|
"grad_norm": 3.9364311500151214,
|
|
"learning_rate": 4.491324795060491e-06,
|
|
"loss": 0.3419,
|
|
"step": 113
|
|
},
|
|
{
|
|
"epoch": 2.425531914893617,
|
|
"grad_norm": 6.072764335330948,
|
|
"learning_rate": 4.4807241083879774e-06,
|
|
"loss": 0.3447,
|
|
"step": 114
|
|
},
|
|
{
|
|
"epoch": 2.4468085106382977,
|
|
"grad_norm": 4.690579387500509,
|
|
"learning_rate": 4.470026884016805e-06,
|
|
"loss": 0.3296,
|
|
"step": 115
|
|
},
|
|
{
|
|
"epoch": 2.4680851063829787,
|
|
"grad_norm": 4.772781283406453,
|
|
"learning_rate": 4.4592336433146e-06,
|
|
"loss": 0.3112,
|
|
"step": 116
|
|
},
|
|
{
|
|
"epoch": 2.4893617021276597,
|
|
"grad_norm": 4.222608771767499,
|
|
"learning_rate": 4.448344912328686e-06,
|
|
"loss": 0.3602,
|
|
"step": 117
|
|
},
|
|
{
|
|
"epoch": 2.5106382978723403,
|
|
"grad_norm": 4.806194387046772,
|
|
"learning_rate": 4.437361221760449e-06,
|
|
"loss": 0.3433,
|
|
"step": 118
|
|
},
|
|
{
|
|
"epoch": 2.5319148936170213,
|
|
"grad_norm": 4.610773667593825,
|
|
"learning_rate": 4.426283106939474e-06,
|
|
"loss": 0.3423,
|
|
"step": 119
|
|
},
|
|
{
|
|
"epoch": 2.5531914893617023,
|
|
"grad_norm": 3.9378113503820114,
|
|
"learning_rate": 4.415111107797445e-06,
|
|
"loss": 0.335,
|
|
"step": 120
|
|
},
|
|
{
|
|
"epoch": 2.574468085106383,
|
|
"grad_norm": 4.262317442324841,
|
|
"learning_rate": 4.403845768841842e-06,
|
|
"loss": 0.2964,
|
|
"step": 121
|
|
},
|
|
{
|
|
"epoch": 2.595744680851064,
|
|
"grad_norm": 4.586717288805483,
|
|
"learning_rate": 4.3924876391293915e-06,
|
|
"loss": 0.3503,
|
|
"step": 122
|
|
},
|
|
{
|
|
"epoch": 2.617021276595745,
|
|
"grad_norm": 4.420522314781192,
|
|
"learning_rate": 4.381037272239311e-06,
|
|
"loss": 0.348,
|
|
"step": 123
|
|
},
|
|
{
|
|
"epoch": 2.6382978723404253,
|
|
"grad_norm": 3.6908626258947694,
|
|
"learning_rate": 4.36949522624633e-06,
|
|
"loss": 0.3473,
|
|
"step": 124
|
|
},
|
|
{
|
|
"epoch": 2.6595744680851063,
|
|
"grad_norm": 4.276095484034931,
|
|
"learning_rate": 4.357862063693486e-06,
|
|
"loss": 0.3734,
|
|
"step": 125
|
|
},
|
|
{
|
|
"epoch": 2.6808510638297873,
|
|
"grad_norm": 4.16461417799402,
|
|
"learning_rate": 4.346138351564711e-06,
|
|
"loss": 0.3563,
|
|
"step": 126
|
|
},
|
|
{
|
|
"epoch": 2.702127659574468,
|
|
"grad_norm": 4.106373804902514,
|
|
"learning_rate": 4.334324661257191e-06,
|
|
"loss": 0.2954,
|
|
"step": 127
|
|
},
|
|
{
|
|
"epoch": 2.723404255319149,
|
|
"grad_norm": 4.113950324587281,
|
|
"learning_rate": 4.322421568553529e-06,
|
|
"loss": 0.3443,
|
|
"step": 128
|
|
},
|
|
{
|
|
"epoch": 2.74468085106383,
|
|
"grad_norm": 4.190826822795899,
|
|
"learning_rate": 4.3104296535936695e-06,
|
|
"loss": 0.3395,
|
|
"step": 129
|
|
},
|
|
{
|
|
"epoch": 2.7659574468085104,
|
|
"grad_norm": 4.9676092124733575,
|
|
"learning_rate": 4.2983495008466285e-06,
|
|
"loss": 0.3789,
|
|
"step": 130
|
|
},
|
|
{
|
|
"epoch": 2.7872340425531914,
|
|
"grad_norm": 4.175058791875455,
|
|
"learning_rate": 4.286181699082008e-06,
|
|
"loss": 0.3388,
|
|
"step": 131
|
|
},
|
|
{
|
|
"epoch": 2.8085106382978724,
|
|
"grad_norm": 5.475615243400478,
|
|
"learning_rate": 4.273926841341303e-06,
|
|
"loss": 0.3614,
|
|
"step": 132
|
|
},
|
|
{
|
|
"epoch": 2.829787234042553,
|
|
"grad_norm": 4.132179706281526,
|
|
"learning_rate": 4.261585524908987e-06,
|
|
"loss": 0.3031,
|
|
"step": 133
|
|
},
|
|
{
|
|
"epoch": 2.851063829787234,
|
|
"grad_norm": 4.52680856181224,
|
|
"learning_rate": 4.249158351283414e-06,
|
|
"loss": 0.3202,
|
|
"step": 134
|
|
},
|
|
{
|
|
"epoch": 2.872340425531915,
|
|
"grad_norm": 4.4158856257333206,
|
|
"learning_rate": 4.236645926147493e-06,
|
|
"loss": 0.3817,
|
|
"step": 135
|
|
},
|
|
{
|
|
"epoch": 2.8936170212765955,
|
|
"grad_norm": 4.131675629353336,
|
|
"learning_rate": 4.224048859339175e-06,
|
|
"loss": 0.325,
|
|
"step": 136
|
|
},
|
|
{
|
|
"epoch": 2.9148936170212765,
|
|
"grad_norm": 4.632609969168001,
|
|
"learning_rate": 4.211367764821722e-06,
|
|
"loss": 0.339,
|
|
"step": 137
|
|
},
|
|
{
|
|
"epoch": 2.9361702127659575,
|
|
"grad_norm": 4.238399617644217,
|
|
"learning_rate": 4.198603260653792e-06,
|
|
"loss": 0.3803,
|
|
"step": 138
|
|
},
|
|
{
|
|
"epoch": 2.9574468085106385,
|
|
"grad_norm": 4.990761801935908,
|
|
"learning_rate": 4.185755968959308e-06,
|
|
"loss": 0.3447,
|
|
"step": 139
|
|
},
|
|
{
|
|
"epoch": 2.978723404255319,
|
|
"grad_norm": 4.548241919258143,
|
|
"learning_rate": 4.172826515897146e-06,
|
|
"loss": 0.3518,
|
|
"step": 140
|
|
},
|
|
{
|
|
"epoch": 3.0,
|
|
"grad_norm": 3.430910087137401,
|
|
"learning_rate": 4.159815531630604e-06,
|
|
"loss": 0.2304,
|
|
"step": 141
|
|
},
|
|
{
|
|
"epoch": 3.021276595744681,
|
|
"grad_norm": 3.668126527915725,
|
|
"learning_rate": 4.146723650296701e-06,
|
|
"loss": 0.2235,
|
|
"step": 142
|
|
},
|
|
{
|
|
"epoch": 3.0425531914893615,
|
|
"grad_norm": 2.8667931520747327,
|
|
"learning_rate": 4.133551509975264e-06,
|
|
"loss": 0.1883,
|
|
"step": 143
|
|
},
|
|
{
|
|
"epoch": 3.0638297872340425,
|
|
"grad_norm": 3.5239577403765106,
|
|
"learning_rate": 4.120299752657828e-06,
|
|
"loss": 0.2217,
|
|
"step": 144
|
|
},
|
|
{
|
|
"epoch": 3.0851063829787235,
|
|
"grad_norm": 3.6921907938710308,
|
|
"learning_rate": 4.106969024216348e-06,
|
|
"loss": 0.2049,
|
|
"step": 145
|
|
},
|
|
{
|
|
"epoch": 3.106382978723404,
|
|
"grad_norm": 3.6392973250804577,
|
|
"learning_rate": 4.093559974371725e-06,
|
|
"loss": 0.1831,
|
|
"step": 146
|
|
},
|
|
{
|
|
"epoch": 3.127659574468085,
|
|
"grad_norm": 3.7365922708313133,
|
|
"learning_rate": 4.080073256662128e-06,
|
|
"loss": 0.198,
|
|
"step": 147
|
|
},
|
|
{
|
|
"epoch": 3.148936170212766,
|
|
"grad_norm": 4.42100805526352,
|
|
"learning_rate": 4.066509528411151e-06,
|
|
"loss": 0.1593,
|
|
"step": 148
|
|
},
|
|
{
|
|
"epoch": 3.1702127659574466,
|
|
"grad_norm": 3.701129708479829,
|
|
"learning_rate": 4.052869450695776e-06,
|
|
"loss": 0.1542,
|
|
"step": 149
|
|
},
|
|
{
|
|
"epoch": 3.1914893617021276,
|
|
"grad_norm": 3.9620190917388793,
|
|
"learning_rate": 4.039153688314146e-06,
|
|
"loss": 0.1673,
|
|
"step": 150
|
|
},
|
|
{
|
|
"epoch": 3.2127659574468086,
|
|
"grad_norm": 4.596659093022736,
|
|
"learning_rate": 4.02536290975317e-06,
|
|
"loss": 0.1781,
|
|
"step": 151
|
|
},
|
|
{
|
|
"epoch": 3.2340425531914896,
|
|
"grad_norm": 4.041018040517897,
|
|
"learning_rate": 4.011497787155938e-06,
|
|
"loss": 0.204,
|
|
"step": 152
|
|
},
|
|
{
|
|
"epoch": 3.25531914893617,
|
|
"grad_norm": 4.4750076668231324,
|
|
"learning_rate": 3.997558996288965e-06,
|
|
"loss": 0.2042,
|
|
"step": 153
|
|
},
|
|
{
|
|
"epoch": 3.276595744680851,
|
|
"grad_norm": 4.328565873011614,
|
|
"learning_rate": 3.983547216509254e-06,
|
|
"loss": 0.206,
|
|
"step": 154
|
|
},
|
|
{
|
|
"epoch": 3.297872340425532,
|
|
"grad_norm": 4.950916510385211,
|
|
"learning_rate": 3.969463130731183e-06,
|
|
"loss": 0.1735,
|
|
"step": 155
|
|
},
|
|
{
|
|
"epoch": 3.3191489361702127,
|
|
"grad_norm": 3.6634515919781356,
|
|
"learning_rate": 3.955307425393224e-06,
|
|
"loss": 0.2152,
|
|
"step": 156
|
|
},
|
|
{
|
|
"epoch": 3.3404255319148937,
|
|
"grad_norm": 3.5183557309327127,
|
|
"learning_rate": 3.941080790424483e-06,
|
|
"loss": 0.1788,
|
|
"step": 157
|
|
},
|
|
{
|
|
"epoch": 3.3617021276595747,
|
|
"grad_norm": 4.2312437308996245,
|
|
"learning_rate": 3.92678391921108e-06,
|
|
"loss": 0.1817,
|
|
"step": 158
|
|
},
|
|
{
|
|
"epoch": 3.382978723404255,
|
|
"grad_norm": 3.4224234758888654,
|
|
"learning_rate": 3.912417508562345e-06,
|
|
"loss": 0.1577,
|
|
"step": 159
|
|
},
|
|
{
|
|
"epoch": 3.404255319148936,
|
|
"grad_norm": 3.7251058715223597,
|
|
"learning_rate": 3.897982258676867e-06,
|
|
"loss": 0.2067,
|
|
"step": 160
|
|
},
|
|
{
|
|
"epoch": 3.425531914893617,
|
|
"grad_norm": 3.123331948611014,
|
|
"learning_rate": 3.88347887310836e-06,
|
|
"loss": 0.1867,
|
|
"step": 161
|
|
},
|
|
{
|
|
"epoch": 3.4468085106382977,
|
|
"grad_norm": 3.040022180599425,
|
|
"learning_rate": 3.868908058731376e-06,
|
|
"loss": 0.1488,
|
|
"step": 162
|
|
},
|
|
{
|
|
"epoch": 3.4680851063829787,
|
|
"grad_norm": 3.936251625166522,
|
|
"learning_rate": 3.85427052570685e-06,
|
|
"loss": 0.175,
|
|
"step": 163
|
|
},
|
|
{
|
|
"epoch": 3.4893617021276597,
|
|
"grad_norm": 3.35819244249543,
|
|
"learning_rate": 3.839566987447492e-06,
|
|
"loss": 0.1992,
|
|
"step": 164
|
|
},
|
|
{
|
|
"epoch": 3.5106382978723403,
|
|
"grad_norm": 3.8098483118828335,
|
|
"learning_rate": 3.824798160583012e-06,
|
|
"loss": 0.1677,
|
|
"step": 165
|
|
},
|
|
{
|
|
"epoch": 3.5319148936170213,
|
|
"grad_norm": 4.08720146453336,
|
|
"learning_rate": 3.8099647649251984e-06,
|
|
"loss": 0.1975,
|
|
"step": 166
|
|
},
|
|
{
|
|
"epoch": 3.5531914893617023,
|
|
"grad_norm": 3.546201338791242,
|
|
"learning_rate": 3.795067523432826e-06,
|
|
"loss": 0.1643,
|
|
"step": 167
|
|
},
|
|
{
|
|
"epoch": 3.574468085106383,
|
|
"grad_norm": 4.022986409946214,
|
|
"learning_rate": 3.780107162176429e-06,
|
|
"loss": 0.2025,
|
|
"step": 168
|
|
},
|
|
{
|
|
"epoch": 3.595744680851064,
|
|
"grad_norm": 4.116301951633153,
|
|
"learning_rate": 3.7650844103029093e-06,
|
|
"loss": 0.2063,
|
|
"step": 169
|
|
},
|
|
{
|
|
"epoch": 3.617021276595745,
|
|
"grad_norm": 5.498041562915274,
|
|
"learning_rate": 3.7500000000000005e-06,
|
|
"loss": 0.1969,
|
|
"step": 170
|
|
},
|
|
{
|
|
"epoch": 3.6382978723404253,
|
|
"grad_norm": 3.571990562222316,
|
|
"learning_rate": 3.7348546664605777e-06,
|
|
"loss": 0.1946,
|
|
"step": 171
|
|
},
|
|
{
|
|
"epoch": 3.6595744680851063,
|
|
"grad_norm": 4.110746067790737,
|
|
"learning_rate": 3.7196491478468322e-06,
|
|
"loss": 0.192,
|
|
"step": 172
|
|
},
|
|
{
|
|
"epoch": 3.6808510638297873,
|
|
"grad_norm": 3.9747616004152984,
|
|
"learning_rate": 3.7043841852542884e-06,
|
|
"loss": 0.2076,
|
|
"step": 173
|
|
},
|
|
{
|
|
"epoch": 3.702127659574468,
|
|
"grad_norm": 4.049886537167864,
|
|
"learning_rate": 3.689060522675689e-06,
|
|
"loss": 0.1885,
|
|
"step": 174
|
|
},
|
|
{
|
|
"epoch": 3.723404255319149,
|
|
"grad_norm": 3.7605106851151593,
|
|
"learning_rate": 3.6736789069647273e-06,
|
|
"loss": 0.1819,
|
|
"step": 175
|
|
},
|
|
{
|
|
"epoch": 3.74468085106383,
|
|
"grad_norm": 4.293344674761706,
|
|
"learning_rate": 3.658240087799655e-06,
|
|
"loss": 0.1782,
|
|
"step": 176
|
|
},
|
|
{
|
|
"epoch": 3.7659574468085104,
|
|
"grad_norm": 3.094426165785156,
|
|
"learning_rate": 3.642744817646736e-06,
|
|
"loss": 0.1562,
|
|
"step": 177
|
|
},
|
|
{
|
|
"epoch": 3.7872340425531914,
|
|
"grad_norm": 4.471233004565249,
|
|
"learning_rate": 3.627193851723577e-06,
|
|
"loss": 0.1538,
|
|
"step": 178
|
|
},
|
|
{
|
|
"epoch": 3.8085106382978724,
|
|
"grad_norm": 3.8346276158744823,
|
|
"learning_rate": 3.611587947962319e-06,
|
|
"loss": 0.2037,
|
|
"step": 179
|
|
},
|
|
{
|
|
"epoch": 3.829787234042553,
|
|
"grad_norm": 4.6309266206267745,
|
|
"learning_rate": 3.595927866972694e-06,
|
|
"loss": 0.1706,
|
|
"step": 180
|
|
},
|
|
{
|
|
"epoch": 3.851063829787234,
|
|
"grad_norm": 3.574650150496055,
|
|
"learning_rate": 3.5802143720049565e-06,
|
|
"loss": 0.1793,
|
|
"step": 181
|
|
},
|
|
{
|
|
"epoch": 3.872340425531915,
|
|
"grad_norm": 4.104316685959246,
|
|
"learning_rate": 3.564448228912682e-06,
|
|
"loss": 0.1997,
|
|
"step": 182
|
|
},
|
|
{
|
|
"epoch": 3.8936170212765955,
|
|
"grad_norm": 3.7077174376752815,
|
|
"learning_rate": 3.5486302061154433e-06,
|
|
"loss": 0.2397,
|
|
"step": 183
|
|
},
|
|
{
|
|
"epoch": 3.9148936170212765,
|
|
"grad_norm": 3.8535618259715863,
|
|
"learning_rate": 3.532761074561355e-06,
|
|
"loss": 0.1724,
|
|
"step": 184
|
|
},
|
|
{
|
|
"epoch": 3.9361702127659575,
|
|
"grad_norm": 4.153454591162772,
|
|
"learning_rate": 3.516841607689501e-06,
|
|
"loss": 0.1879,
|
|
"step": 185
|
|
},
|
|
{
|
|
"epoch": 3.9574468085106385,
|
|
"grad_norm": 3.2186742898572254,
|
|
"learning_rate": 3.5008725813922383e-06,
|
|
"loss": 0.1793,
|
|
"step": 186
|
|
},
|
|
{
|
|
"epoch": 3.978723404255319,
|
|
"grad_norm": 3.8852555348009066,
|
|
"learning_rate": 3.4848547739773782e-06,
|
|
"loss": 0.2205,
|
|
"step": 187
|
|
},
|
|
{
|
|
"epoch": 4.0,
|
|
"grad_norm": 3.1773700849249984,
|
|
"learning_rate": 3.4687889661302577e-06,
|
|
"loss": 0.1375,
|
|
"step": 188
|
|
},
|
|
{
|
|
"epoch": 4.0212765957446805,
|
|
"grad_norm": 2.909112047920458,
|
|
"learning_rate": 3.452675940875686e-06,
|
|
"loss": 0.1081,
|
|
"step": 189
|
|
},
|
|
{
|
|
"epoch": 4.042553191489362,
|
|
"grad_norm": 2.5683738553439794,
|
|
"learning_rate": 3.436516483539781e-06,
|
|
"loss": 0.0953,
|
|
"step": 190
|
|
},
|
|
{
|
|
"epoch": 4.0638297872340425,
|
|
"grad_norm": 2.9198721500124845,
|
|
"learning_rate": 3.4203113817116955e-06,
|
|
"loss": 0.1053,
|
|
"step": 191
|
|
},
|
|
{
|
|
"epoch": 4.085106382978723,
|
|
"grad_norm": 2.24010493827577,
|
|
"learning_rate": 3.4040614252052305e-06,
|
|
"loss": 0.0927,
|
|
"step": 192
|
|
},
|
|
{
|
|
"epoch": 4.1063829787234045,
|
|
"grad_norm": 2.878670182658427,
|
|
"learning_rate": 3.387767406020343e-06,
|
|
"loss": 0.103,
|
|
"step": 193
|
|
},
|
|
{
|
|
"epoch": 4.127659574468085,
|
|
"grad_norm": 3.1788187649976978,
|
|
"learning_rate": 3.3714301183045382e-06,
|
|
"loss": 0.1118,
|
|
"step": 194
|
|
},
|
|
{
|
|
"epoch": 4.148936170212766,
|
|
"grad_norm": 3.3067413198545634,
|
|
"learning_rate": 3.3550503583141726e-06,
|
|
"loss": 0.1099,
|
|
"step": 195
|
|
},
|
|
{
|
|
"epoch": 4.170212765957447,
|
|
"grad_norm": 2.319050087976373,
|
|
"learning_rate": 3.338628924375638e-06,
|
|
"loss": 0.0898,
|
|
"step": 196
|
|
},
|
|
{
|
|
"epoch": 4.191489361702128,
|
|
"grad_norm": 2.6474547453865895,
|
|
"learning_rate": 3.3221666168464584e-06,
|
|
"loss": 0.0915,
|
|
"step": 197
|
|
},
|
|
{
|
|
"epoch": 4.212765957446808,
|
|
"grad_norm": 3.2898592433873217,
|
|
"learning_rate": 3.3056642380762783e-06,
|
|
"loss": 0.076,
|
|
"step": 198
|
|
},
|
|
{
|
|
"epoch": 4.23404255319149,
|
|
"grad_norm": 3.7084220696956685,
|
|
"learning_rate": 3.2891225923677565e-06,
|
|
"loss": 0.0832,
|
|
"step": 199
|
|
},
|
|
{
|
|
"epoch": 4.25531914893617,
|
|
"grad_norm": 3.853474111497437,
|
|
"learning_rate": 3.272542485937369e-06,
|
|
"loss": 0.1113,
|
|
"step": 200
|
|
},
|
|
{
|
|
"epoch": 4.25531914893617,
|
|
"eval_loss": 1.3551427125930786,
|
|
"eval_runtime": 1.3449,
|
|
"eval_samples_per_second": 17.102,
|
|
"eval_steps_per_second": 0.744,
|
|
"step": 200
|
|
},
|
|
{
|
|
"epoch": 4.276595744680851,
|
|
"grad_norm": 3.1480732576505686,
|
|
"learning_rate": 3.2559247268761117e-06,
|
|
"loss": 0.0894,
|
|
"step": 201
|
|
},
|
|
{
|
|
"epoch": 4.297872340425532,
|
|
"grad_norm": 4.224708808728799,
|
|
"learning_rate": 3.2392701251101172e-06,
|
|
"loss": 0.1037,
|
|
"step": 202
|
|
},
|
|
{
|
|
"epoch": 4.319148936170213,
|
|
"grad_norm": 3.127550241031144,
|
|
"learning_rate": 3.222579492361179e-06,
|
|
"loss": 0.1193,
|
|
"step": 203
|
|
},
|
|
{
|
|
"epoch": 4.340425531914893,
|
|
"grad_norm": 3.358278204814404,
|
|
"learning_rate": 3.205853642107192e-06,
|
|
"loss": 0.1088,
|
|
"step": 204
|
|
},
|
|
{
|
|
"epoch": 4.361702127659575,
|
|
"grad_norm": 3.5723872992052668,
|
|
"learning_rate": 3.189093389542498e-06,
|
|
"loss": 0.1156,
|
|
"step": 205
|
|
},
|
|
{
|
|
"epoch": 4.382978723404255,
|
|
"grad_norm": 3.3910498011991863,
|
|
"learning_rate": 3.1722995515381644e-06,
|
|
"loss": 0.0813,
|
|
"step": 206
|
|
},
|
|
{
|
|
"epoch": 4.404255319148936,
|
|
"grad_norm": 3.987898875106463,
|
|
"learning_rate": 3.155472946602162e-06,
|
|
"loss": 0.1063,
|
|
"step": 207
|
|
},
|
|
{
|
|
"epoch": 4.425531914893617,
|
|
"grad_norm": 3.0288381078835416,
|
|
"learning_rate": 3.1386143948394764e-06,
|
|
"loss": 0.1089,
|
|
"step": 208
|
|
},
|
|
{
|
|
"epoch": 4.446808510638298,
|
|
"grad_norm": 3.3307094726424635,
|
|
"learning_rate": 3.121724717912138e-06,
|
|
"loss": 0.1142,
|
|
"step": 209
|
|
},
|
|
{
|
|
"epoch": 4.468085106382979,
|
|
"grad_norm": 2.9848172012011664,
|
|
"learning_rate": 3.1048047389991693e-06,
|
|
"loss": 0.1157,
|
|
"step": 210
|
|
},
|
|
{
|
|
"epoch": 4.48936170212766,
|
|
"grad_norm": 2.978438802813703,
|
|
"learning_rate": 3.087855282756475e-06,
|
|
"loss": 0.0942,
|
|
"step": 211
|
|
},
|
|
{
|
|
"epoch": 4.51063829787234,
|
|
"grad_norm": 2.8670301514814565,
|
|
"learning_rate": 3.0708771752766397e-06,
|
|
"loss": 0.0876,
|
|
"step": 212
|
|
},
|
|
{
|
|
"epoch": 4.531914893617021,
|
|
"grad_norm": 3.0160259648937626,
|
|
"learning_rate": 3.053871244048669e-06,
|
|
"loss": 0.1057,
|
|
"step": 213
|
|
},
|
|
{
|
|
"epoch": 4.553191489361702,
|
|
"grad_norm": 3.552730667163276,
|
|
"learning_rate": 3.0368383179176584e-06,
|
|
"loss": 0.0681,
|
|
"step": 214
|
|
},
|
|
{
|
|
"epoch": 4.574468085106383,
|
|
"grad_norm": 3.0671796344169744,
|
|
"learning_rate": 3.019779227044398e-06,
|
|
"loss": 0.1052,
|
|
"step": 215
|
|
},
|
|
{
|
|
"epoch": 4.595744680851064,
|
|
"grad_norm": 3.4917736614570343,
|
|
"learning_rate": 3.002694802864912e-06,
|
|
"loss": 0.1131,
|
|
"step": 216
|
|
},
|
|
{
|
|
"epoch": 4.617021276595745,
|
|
"grad_norm": 2.997068423954694,
|
|
"learning_rate": 2.98558587804993e-06,
|
|
"loss": 0.1249,
|
|
"step": 217
|
|
},
|
|
{
|
|
"epoch": 4.638297872340425,
|
|
"grad_norm": 2.945598337302224,
|
|
"learning_rate": 2.9684532864643123e-06,
|
|
"loss": 0.0916,
|
|
"step": 218
|
|
},
|
|
{
|
|
"epoch": 4.659574468085106,
|
|
"grad_norm": 2.5534208875346494,
|
|
"learning_rate": 2.9512978631264006e-06,
|
|
"loss": 0.0834,
|
|
"step": 219
|
|
},
|
|
{
|
|
"epoch": 4.680851063829787,
|
|
"grad_norm": 2.3947639549989144,
|
|
"learning_rate": 2.9341204441673267e-06,
|
|
"loss": 0.0974,
|
|
"step": 220
|
|
},
|
|
{
|
|
"epoch": 4.702127659574468,
|
|
"grad_norm": 3.3251173109542074,
|
|
"learning_rate": 2.9169218667902562e-06,
|
|
"loss": 0.1011,
|
|
"step": 221
|
|
},
|
|
{
|
|
"epoch": 4.723404255319149,
|
|
"grad_norm": 3.059268693430171,
|
|
"learning_rate": 2.8997029692295875e-06,
|
|
"loss": 0.0821,
|
|
"step": 222
|
|
},
|
|
{
|
|
"epoch": 4.74468085106383,
|
|
"grad_norm": 3.5597698182253965,
|
|
"learning_rate": 2.8824645907100957e-06,
|
|
"loss": 0.0856,
|
|
"step": 223
|
|
},
|
|
{
|
|
"epoch": 4.76595744680851,
|
|
"grad_norm": 3.3566578614960423,
|
|
"learning_rate": 2.8652075714060296e-06,
|
|
"loss": 0.132,
|
|
"step": 224
|
|
},
|
|
{
|
|
"epoch": 4.787234042553192,
|
|
"grad_norm": 3.470592088594595,
|
|
"learning_rate": 2.847932752400164e-06,
|
|
"loss": 0.0958,
|
|
"step": 225
|
|
},
|
|
{
|
|
"epoch": 4.808510638297872,
|
|
"grad_norm": 2.911346797843613,
|
|
"learning_rate": 2.8306409756428067e-06,
|
|
"loss": 0.0894,
|
|
"step": 226
|
|
},
|
|
{
|
|
"epoch": 4.829787234042553,
|
|
"grad_norm": 4.613111358505509,
|
|
"learning_rate": 2.813333083910761e-06,
|
|
"loss": 0.1166,
|
|
"step": 227
|
|
},
|
|
{
|
|
"epoch": 4.851063829787234,
|
|
"grad_norm": 3.015962228404618,
|
|
"learning_rate": 2.7960099207662535e-06,
|
|
"loss": 0.1155,
|
|
"step": 228
|
|
},
|
|
{
|
|
"epoch": 4.872340425531915,
|
|
"grad_norm": 2.987399829854267,
|
|
"learning_rate": 2.778672330515814e-06,
|
|
"loss": 0.1104,
|
|
"step": 229
|
|
},
|
|
{
|
|
"epoch": 4.8936170212765955,
|
|
"grad_norm": 2.829987084084442,
|
|
"learning_rate": 2.761321158169134e-06,
|
|
"loss": 0.1066,
|
|
"step": 230
|
|
},
|
|
{
|
|
"epoch": 4.914893617021277,
|
|
"grad_norm": 3.708679514606222,
|
|
"learning_rate": 2.743957249397874e-06,
|
|
"loss": 0.1048,
|
|
"step": 231
|
|
},
|
|
{
|
|
"epoch": 4.9361702127659575,
|
|
"grad_norm": 2.749591872522463,
|
|
"learning_rate": 2.726581450494451e-06,
|
|
"loss": 0.1159,
|
|
"step": 232
|
|
},
|
|
{
|
|
"epoch": 4.957446808510638,
|
|
"grad_norm": 3.776661504783798,
|
|
"learning_rate": 2.70919460833079e-06,
|
|
"loss": 0.1167,
|
|
"step": 233
|
|
},
|
|
{
|
|
"epoch": 4.9787234042553195,
|
|
"grad_norm": 3.091697194785843,
|
|
"learning_rate": 2.6917975703170466e-06,
|
|
"loss": 0.1174,
|
|
"step": 234
|
|
},
|
|
{
|
|
"epoch": 5.0,
|
|
"grad_norm": 2.8077659070563192,
|
|
"learning_rate": 2.6743911843603134e-06,
|
|
"loss": 0.0737,
|
|
"step": 235
|
|
},
|
|
{
|
|
"epoch": 5.0212765957446805,
|
|
"grad_norm": 1.9864860542138973,
|
|
"learning_rate": 2.6569762988232838e-06,
|
|
"loss": 0.0498,
|
|
"step": 236
|
|
},
|
|
{
|
|
"epoch": 5.042553191489362,
|
|
"grad_norm": 2.080148938371656,
|
|
"learning_rate": 2.63955376248291e-06,
|
|
"loss": 0.0588,
|
|
"step": 237
|
|
},
|
|
{
|
|
"epoch": 5.0638297872340425,
|
|
"grad_norm": 2.4271830087344215,
|
|
"learning_rate": 2.6221244244890336e-06,
|
|
"loss": 0.04,
|
|
"step": 238
|
|
},
|
|
{
|
|
"epoch": 5.085106382978723,
|
|
"grad_norm": 2.182065583755821,
|
|
"learning_rate": 2.604689134322999e-06,
|
|
"loss": 0.0635,
|
|
"step": 239
|
|
},
|
|
{
|
|
"epoch": 5.1063829787234045,
|
|
"grad_norm": 2.3865912853026727,
|
|
"learning_rate": 2.587248741756253e-06,
|
|
"loss": 0.0602,
|
|
"step": 240
|
|
},
|
|
{
|
|
"epoch": 5.127659574468085,
|
|
"grad_norm": 2.3131711827099832,
|
|
"learning_rate": 2.569804096808923e-06,
|
|
"loss": 0.0709,
|
|
"step": 241
|
|
},
|
|
{
|
|
"epoch": 5.148936170212766,
|
|
"grad_norm": 2.3050528514219004,
|
|
"learning_rate": 2.5523560497083927e-06,
|
|
"loss": 0.0547,
|
|
"step": 242
|
|
},
|
|
{
|
|
"epoch": 5.170212765957447,
|
|
"grad_norm": 2.124162166492525,
|
|
"learning_rate": 2.5349054508478636e-06,
|
|
"loss": 0.046,
|
|
"step": 243
|
|
},
|
|
{
|
|
"epoch": 5.191489361702128,
|
|
"grad_norm": 1.9773286780066965,
|
|
"learning_rate": 2.517453150744904e-06,
|
|
"loss": 0.0432,
|
|
"step": 244
|
|
},
|
|
{
|
|
"epoch": 5.212765957446808,
|
|
"grad_norm": 3.116678592316818,
|
|
"learning_rate": 2.5e-06,
|
|
"loss": 0.0634,
|
|
"step": 245
|
|
},
|
|
{
|
|
"epoch": 5.23404255319149,
|
|
"grad_norm": 2.504016738413573,
|
|
"learning_rate": 2.482546849255096e-06,
|
|
"loss": 0.0589,
|
|
"step": 246
|
|
},
|
|
{
|
|
"epoch": 5.25531914893617,
|
|
"grad_norm": 3.505324999262586,
|
|
"learning_rate": 2.4650945491521372e-06,
|
|
"loss": 0.0547,
|
|
"step": 247
|
|
},
|
|
{
|
|
"epoch": 5.276595744680851,
|
|
"grad_norm": 2.2697348908750974,
|
|
"learning_rate": 2.447643950291608e-06,
|
|
"loss": 0.0608,
|
|
"step": 248
|
|
},
|
|
{
|
|
"epoch": 5.297872340425532,
|
|
"grad_norm": 2.8775304018384986,
|
|
"learning_rate": 2.4301959031910785e-06,
|
|
"loss": 0.0765,
|
|
"step": 249
|
|
},
|
|
{
|
|
"epoch": 5.319148936170213,
|
|
"grad_norm": 2.4093465444498614,
|
|
"learning_rate": 2.4127512582437486e-06,
|
|
"loss": 0.0583,
|
|
"step": 250
|
|
},
|
|
{
|
|
"epoch": 5.340425531914893,
|
|
"grad_norm": 2.365667376680801,
|
|
"learning_rate": 2.3953108656770018e-06,
|
|
"loss": 0.0629,
|
|
"step": 251
|
|
},
|
|
{
|
|
"epoch": 5.361702127659575,
|
|
"grad_norm": 2.713559880628167,
|
|
"learning_rate": 2.377875575510967e-06,
|
|
"loss": 0.0669,
|
|
"step": 252
|
|
},
|
|
{
|
|
"epoch": 5.382978723404255,
|
|
"grad_norm": 2.999000225951145,
|
|
"learning_rate": 2.3604462375170905e-06,
|
|
"loss": 0.0617,
|
|
"step": 253
|
|
},
|
|
{
|
|
"epoch": 5.404255319148936,
|
|
"grad_norm": 2.6164170571843512,
|
|
"learning_rate": 2.3430237011767166e-06,
|
|
"loss": 0.0582,
|
|
"step": 254
|
|
},
|
|
{
|
|
"epoch": 5.425531914893617,
|
|
"grad_norm": 2.4039871330574796,
|
|
"learning_rate": 2.325608815639687e-06,
|
|
"loss": 0.0578,
|
|
"step": 255
|
|
},
|
|
{
|
|
"epoch": 5.446808510638298,
|
|
"grad_norm": 2.97747740854911,
|
|
"learning_rate": 2.3082024296829538e-06,
|
|
"loss": 0.0659,
|
|
"step": 256
|
|
},
|
|
{
|
|
"epoch": 5.468085106382979,
|
|
"grad_norm": 2.336759542261552,
|
|
"learning_rate": 2.290805391669212e-06,
|
|
"loss": 0.0548,
|
|
"step": 257
|
|
},
|
|
{
|
|
"epoch": 5.48936170212766,
|
|
"grad_norm": 3.1670650790685846,
|
|
"learning_rate": 2.2734185495055503e-06,
|
|
"loss": 0.0585,
|
|
"step": 258
|
|
},
|
|
{
|
|
"epoch": 5.51063829787234,
|
|
"grad_norm": 3.2034707124777952,
|
|
"learning_rate": 2.256042750602127e-06,
|
|
"loss": 0.0593,
|
|
"step": 259
|
|
},
|
|
{
|
|
"epoch": 5.531914893617021,
|
|
"grad_norm": 2.3348070175379503,
|
|
"learning_rate": 2.238678841830867e-06,
|
|
"loss": 0.0547,
|
|
"step": 260
|
|
},
|
|
{
|
|
"epoch": 5.553191489361702,
|
|
"grad_norm": 2.608427512535968,
|
|
"learning_rate": 2.2213276694841866e-06,
|
|
"loss": 0.0643,
|
|
"step": 261
|
|
},
|
|
{
|
|
"epoch": 5.574468085106383,
|
|
"grad_norm": 2.4848496472118495,
|
|
"learning_rate": 2.2039900792337477e-06,
|
|
"loss": 0.0608,
|
|
"step": 262
|
|
},
|
|
{
|
|
"epoch": 5.595744680851064,
|
|
"grad_norm": 1.886510569911626,
|
|
"learning_rate": 2.186666916089239e-06,
|
|
"loss": 0.0599,
|
|
"step": 263
|
|
},
|
|
{
|
|
"epoch": 5.617021276595745,
|
|
"grad_norm": 2.8635289633291547,
|
|
"learning_rate": 2.1693590243571937e-06,
|
|
"loss": 0.0515,
|
|
"step": 264
|
|
},
|
|
{
|
|
"epoch": 5.638297872340425,
|
|
"grad_norm": 2.854324373551202,
|
|
"learning_rate": 2.1520672475998374e-06,
|
|
"loss": 0.0554,
|
|
"step": 265
|
|
},
|
|
{
|
|
"epoch": 5.659574468085106,
|
|
"grad_norm": 1.8544197687406436,
|
|
"learning_rate": 2.134792428593971e-06,
|
|
"loss": 0.0421,
|
|
"step": 266
|
|
},
|
|
{
|
|
"epoch": 5.680851063829787,
|
|
"grad_norm": 2.6205983731152322,
|
|
"learning_rate": 2.117535409289905e-06,
|
|
"loss": 0.0457,
|
|
"step": 267
|
|
},
|
|
{
|
|
"epoch": 5.702127659574468,
|
|
"grad_norm": 1.7630228804740633,
|
|
"learning_rate": 2.1002970307704134e-06,
|
|
"loss": 0.0369,
|
|
"step": 268
|
|
},
|
|
{
|
|
"epoch": 5.723404255319149,
|
|
"grad_norm": 2.638068560108755,
|
|
"learning_rate": 2.0830781332097446e-06,
|
|
"loss": 0.0583,
|
|
"step": 269
|
|
},
|
|
{
|
|
"epoch": 5.74468085106383,
|
|
"grad_norm": 3.1753027345645646,
|
|
"learning_rate": 2.0658795558326745e-06,
|
|
"loss": 0.0674,
|
|
"step": 270
|
|
},
|
|
{
|
|
"epoch": 5.76595744680851,
|
|
"grad_norm": 2.2642667997058257,
|
|
"learning_rate": 2.0487021368736002e-06,
|
|
"loss": 0.064,
|
|
"step": 271
|
|
},
|
|
{
|
|
"epoch": 5.787234042553192,
|
|
"grad_norm": 2.559426313332899,
|
|
"learning_rate": 2.031546713535688e-06,
|
|
"loss": 0.0549,
|
|
"step": 272
|
|
},
|
|
{
|
|
"epoch": 5.808510638297872,
|
|
"grad_norm": 2.878855059930482,
|
|
"learning_rate": 2.0144141219500707e-06,
|
|
"loss": 0.0559,
|
|
"step": 273
|
|
},
|
|
{
|
|
"epoch": 5.829787234042553,
|
|
"grad_norm": 2.283778113411308,
|
|
"learning_rate": 1.997305197135089e-06,
|
|
"loss": 0.0548,
|
|
"step": 274
|
|
},
|
|
{
|
|
"epoch": 5.851063829787234,
|
|
"grad_norm": 2.4192051614394003,
|
|
"learning_rate": 1.9802207729556023e-06,
|
|
"loss": 0.0502,
|
|
"step": 275
|
|
},
|
|
{
|
|
"epoch": 5.872340425531915,
|
|
"grad_norm": 2.322479824517103,
|
|
"learning_rate": 1.963161682082342e-06,
|
|
"loss": 0.064,
|
|
"step": 276
|
|
},
|
|
{
|
|
"epoch": 5.8936170212765955,
|
|
"grad_norm": 2.4593833670456005,
|
|
"learning_rate": 1.946128755951332e-06,
|
|
"loss": 0.058,
|
|
"step": 277
|
|
},
|
|
{
|
|
"epoch": 5.914893617021277,
|
|
"grad_norm": 2.5251545274407907,
|
|
"learning_rate": 1.9291228247233607e-06,
|
|
"loss": 0.0522,
|
|
"step": 278
|
|
},
|
|
{
|
|
"epoch": 5.9361702127659575,
|
|
"grad_norm": 1.8104094730869373,
|
|
"learning_rate": 1.912144717243525e-06,
|
|
"loss": 0.0296,
|
|
"step": 279
|
|
},
|
|
{
|
|
"epoch": 5.957446808510638,
|
|
"grad_norm": 2.689146388919388,
|
|
"learning_rate": 1.895195261000831e-06,
|
|
"loss": 0.0407,
|
|
"step": 280
|
|
},
|
|
{
|
|
"epoch": 5.9787234042553195,
|
|
"grad_norm": 1.8669924342723785,
|
|
"learning_rate": 1.8782752820878636e-06,
|
|
"loss": 0.0495,
|
|
"step": 281
|
|
},
|
|
{
|
|
"epoch": 6.0,
|
|
"grad_norm": 2.271539062015242,
|
|
"learning_rate": 1.8613856051605242e-06,
|
|
"loss": 0.0455,
|
|
"step": 282
|
|
},
|
|
{
|
|
"epoch": 6.0212765957446805,
|
|
"grad_norm": 1.525504212144237,
|
|
"learning_rate": 1.8445270533978387e-06,
|
|
"loss": 0.0377,
|
|
"step": 283
|
|
},
|
|
{
|
|
"epoch": 6.042553191489362,
|
|
"grad_norm": 1.0859867252268387,
|
|
"learning_rate": 1.827700448461836e-06,
|
|
"loss": 0.0221,
|
|
"step": 284
|
|
},
|
|
{
|
|
"epoch": 6.0638297872340425,
|
|
"grad_norm": 1.6689218206446002,
|
|
"learning_rate": 1.8109066104575023e-06,
|
|
"loss": 0.0237,
|
|
"step": 285
|
|
},
|
|
{
|
|
"epoch": 6.085106382978723,
|
|
"grad_norm": 1.7091034732477,
|
|
"learning_rate": 1.7941463578928088e-06,
|
|
"loss": 0.0316,
|
|
"step": 286
|
|
},
|
|
{
|
|
"epoch": 6.1063829787234045,
|
|
"grad_norm": 1.473946493341099,
|
|
"learning_rate": 1.7774205076388207e-06,
|
|
"loss": 0.0301,
|
|
"step": 287
|
|
},
|
|
{
|
|
"epoch": 6.127659574468085,
|
|
"grad_norm": 1.9210720976922275,
|
|
"learning_rate": 1.7607298748898844e-06,
|
|
"loss": 0.0241,
|
|
"step": 288
|
|
},
|
|
{
|
|
"epoch": 6.148936170212766,
|
|
"grad_norm": 2.2803503758262504,
|
|
"learning_rate": 1.744075273123889e-06,
|
|
"loss": 0.0454,
|
|
"step": 289
|
|
},
|
|
{
|
|
"epoch": 6.170212765957447,
|
|
"grad_norm": 2.024197465049683,
|
|
"learning_rate": 1.7274575140626318e-06,
|
|
"loss": 0.0362,
|
|
"step": 290
|
|
},
|
|
{
|
|
"epoch": 6.191489361702128,
|
|
"grad_norm": 1.84667704080016,
|
|
"learning_rate": 1.7108774076322443e-06,
|
|
"loss": 0.0403,
|
|
"step": 291
|
|
},
|
|
{
|
|
"epoch": 6.212765957446808,
|
|
"grad_norm": 1.695502301672247,
|
|
"learning_rate": 1.6943357619237227e-06,
|
|
"loss": 0.0306,
|
|
"step": 292
|
|
},
|
|
{
|
|
"epoch": 6.23404255319149,
|
|
"grad_norm": 1.6806451272904126,
|
|
"learning_rate": 1.677833383153542e-06,
|
|
"loss": 0.0284,
|
|
"step": 293
|
|
},
|
|
{
|
|
"epoch": 6.25531914893617,
|
|
"grad_norm": 2.9118161159346854,
|
|
"learning_rate": 1.661371075624363e-06,
|
|
"loss": 0.0256,
|
|
"step": 294
|
|
},
|
|
{
|
|
"epoch": 6.276595744680851,
|
|
"grad_norm": 1.8173030910949344,
|
|
"learning_rate": 1.6449496416858285e-06,
|
|
"loss": 0.0284,
|
|
"step": 295
|
|
},
|
|
{
|
|
"epoch": 6.297872340425532,
|
|
"grad_norm": 2.223540017311328,
|
|
"learning_rate": 1.6285698816954626e-06,
|
|
"loss": 0.0443,
|
|
"step": 296
|
|
},
|
|
{
|
|
"epoch": 6.319148936170213,
|
|
"grad_norm": 1.4963288286875704,
|
|
"learning_rate": 1.612232593979658e-06,
|
|
"loss": 0.0367,
|
|
"step": 297
|
|
},
|
|
{
|
|
"epoch": 6.340425531914893,
|
|
"grad_norm": 1.3236418660083258,
|
|
"learning_rate": 1.5959385747947697e-06,
|
|
"loss": 0.0168,
|
|
"step": 298
|
|
},
|
|
{
|
|
"epoch": 6.361702127659575,
|
|
"grad_norm": 2.1082580928006704,
|
|
"learning_rate": 1.5796886182883053e-06,
|
|
"loss": 0.0384,
|
|
"step": 299
|
|
},
|
|
{
|
|
"epoch": 6.382978723404255,
|
|
"grad_norm": 2.0461518407784665,
|
|
"learning_rate": 1.56348351646022e-06,
|
|
"loss": 0.0348,
|
|
"step": 300
|
|
},
|
|
{
|
|
"epoch": 6.404255319148936,
|
|
"grad_norm": 1.9027154330548208,
|
|
"learning_rate": 1.547324059124315e-06,
|
|
"loss": 0.0409,
|
|
"step": 301
|
|
},
|
|
{
|
|
"epoch": 6.425531914893617,
|
|
"grad_norm": 1.8430627981587546,
|
|
"learning_rate": 1.5312110338697427e-06,
|
|
"loss": 0.0336,
|
|
"step": 302
|
|
},
|
|
{
|
|
"epoch": 6.446808510638298,
|
|
"grad_norm": 2.013051690553332,
|
|
"learning_rate": 1.5151452260226224e-06,
|
|
"loss": 0.0361,
|
|
"step": 303
|
|
},
|
|
{
|
|
"epoch": 6.468085106382979,
|
|
"grad_norm": 2.1207015485047753,
|
|
"learning_rate": 1.4991274186077632e-06,
|
|
"loss": 0.0405,
|
|
"step": 304
|
|
},
|
|
{
|
|
"epoch": 6.48936170212766,
|
|
"grad_norm": 1.3483418905294124,
|
|
"learning_rate": 1.4831583923105e-06,
|
|
"loss": 0.0264,
|
|
"step": 305
|
|
},
|
|
{
|
|
"epoch": 6.51063829787234,
|
|
"grad_norm": 1.2898003351353924,
|
|
"learning_rate": 1.467238925438646e-06,
|
|
"loss": 0.0326,
|
|
"step": 306
|
|
},
|
|
{
|
|
"epoch": 6.531914893617021,
|
|
"grad_norm": 1.83413124159608,
|
|
"learning_rate": 1.4513697938845571e-06,
|
|
"loss": 0.0329,
|
|
"step": 307
|
|
},
|
|
{
|
|
"epoch": 6.553191489361702,
|
|
"grad_norm": 1.9300831447222762,
|
|
"learning_rate": 1.4355517710873184e-06,
|
|
"loss": 0.0342,
|
|
"step": 308
|
|
},
|
|
{
|
|
"epoch": 6.574468085106383,
|
|
"grad_norm": 2.46498295305994,
|
|
"learning_rate": 1.419785627995044e-06,
|
|
"loss": 0.0293,
|
|
"step": 309
|
|
},
|
|
{
|
|
"epoch": 6.595744680851064,
|
|
"grad_norm": 1.8810365903428496,
|
|
"learning_rate": 1.4040721330273063e-06,
|
|
"loss": 0.0385,
|
|
"step": 310
|
|
},
|
|
{
|
|
"epoch": 6.617021276595745,
|
|
"grad_norm": 1.7341556708700896,
|
|
"learning_rate": 1.388412052037682e-06,
|
|
"loss": 0.0373,
|
|
"step": 311
|
|
},
|
|
{
|
|
"epoch": 6.638297872340425,
|
|
"grad_norm": 1.4886715538593713,
|
|
"learning_rate": 1.3728061482764238e-06,
|
|
"loss": 0.0374,
|
|
"step": 312
|
|
},
|
|
{
|
|
"epoch": 6.659574468085106,
|
|
"grad_norm": 1.8999495842705894,
|
|
"learning_rate": 1.3572551823532654e-06,
|
|
"loss": 0.0316,
|
|
"step": 313
|
|
},
|
|
{
|
|
"epoch": 6.680851063829787,
|
|
"grad_norm": 1.464463066404411,
|
|
"learning_rate": 1.3417599122003464e-06,
|
|
"loss": 0.033,
|
|
"step": 314
|
|
},
|
|
{
|
|
"epoch": 6.702127659574468,
|
|
"grad_norm": 1.8396892711483188,
|
|
"learning_rate": 1.3263210930352737e-06,
|
|
"loss": 0.0351,
|
|
"step": 315
|
|
},
|
|
{
|
|
"epoch": 6.723404255319149,
|
|
"grad_norm": 1.555502571877375,
|
|
"learning_rate": 1.3109394773243117e-06,
|
|
"loss": 0.0333,
|
|
"step": 316
|
|
},
|
|
{
|
|
"epoch": 6.74468085106383,
|
|
"grad_norm": 2.124808645505887,
|
|
"learning_rate": 1.2956158147457116e-06,
|
|
"loss": 0.0294,
|
|
"step": 317
|
|
},
|
|
{
|
|
"epoch": 6.76595744680851,
|
|
"grad_norm": 1.7688895712588661,
|
|
"learning_rate": 1.280350852153168e-06,
|
|
"loss": 0.0242,
|
|
"step": 318
|
|
},
|
|
{
|
|
"epoch": 6.787234042553192,
|
|
"grad_norm": 2.0383431308499325,
|
|
"learning_rate": 1.2651453335394232e-06,
|
|
"loss": 0.0353,
|
|
"step": 319
|
|
},
|
|
{
|
|
"epoch": 6.808510638297872,
|
|
"grad_norm": 1.7645071618559023,
|
|
"learning_rate": 1.2500000000000007e-06,
|
|
"loss": 0.0299,
|
|
"step": 320
|
|
},
|
|
{
|
|
"epoch": 6.829787234042553,
|
|
"grad_norm": 2.0474065053014217,
|
|
"learning_rate": 1.234915589697091e-06,
|
|
"loss": 0.0297,
|
|
"step": 321
|
|
},
|
|
{
|
|
"epoch": 6.851063829787234,
|
|
"grad_norm": 1.7823278045268314,
|
|
"learning_rate": 1.2198928378235717e-06,
|
|
"loss": 0.038,
|
|
"step": 322
|
|
},
|
|
{
|
|
"epoch": 6.872340425531915,
|
|
"grad_norm": 1.2380075941515003,
|
|
"learning_rate": 1.204932476567175e-06,
|
|
"loss": 0.031,
|
|
"step": 323
|
|
},
|
|
{
|
|
"epoch": 6.8936170212765955,
|
|
"grad_norm": 1.9796302395422996,
|
|
"learning_rate": 1.1900352350748026e-06,
|
|
"loss": 0.0381,
|
|
"step": 324
|
|
},
|
|
{
|
|
"epoch": 6.914893617021277,
|
|
"grad_norm": 1.3804831755312008,
|
|
"learning_rate": 1.1752018394169882e-06,
|
|
"loss": 0.0316,
|
|
"step": 325
|
|
},
|
|
{
|
|
"epoch": 6.9361702127659575,
|
|
"grad_norm": 1.6152383977065423,
|
|
"learning_rate": 1.160433012552508e-06,
|
|
"loss": 0.0343,
|
|
"step": 326
|
|
},
|
|
{
|
|
"epoch": 6.957446808510638,
|
|
"grad_norm": 1.69638635732096,
|
|
"learning_rate": 1.1457294742931508e-06,
|
|
"loss": 0.0296,
|
|
"step": 327
|
|
},
|
|
{
|
|
"epoch": 6.9787234042553195,
|
|
"grad_norm": 2.0806576931686456,
|
|
"learning_rate": 1.1310919412686248e-06,
|
|
"loss": 0.0341,
|
|
"step": 328
|
|
},
|
|
{
|
|
"epoch": 7.0,
|
|
"grad_norm": 1.8303598912826444,
|
|
"learning_rate": 1.11652112689164e-06,
|
|
"loss": 0.0405,
|
|
"step": 329
|
|
},
|
|
{
|
|
"epoch": 7.0212765957446805,
|
|
"grad_norm": 1.5859756930815514,
|
|
"learning_rate": 1.1020177413231334e-06,
|
|
"loss": 0.0252,
|
|
"step": 330
|
|
},
|
|
{
|
|
"epoch": 7.042553191489362,
|
|
"grad_norm": 0.7609936522020677,
|
|
"learning_rate": 1.0875824914376555e-06,
|
|
"loss": 0.0158,
|
|
"step": 331
|
|
},
|
|
{
|
|
"epoch": 7.0638297872340425,
|
|
"grad_norm": 2.0790528777967188,
|
|
"learning_rate": 1.073216080788921e-06,
|
|
"loss": 0.0175,
|
|
"step": 332
|
|
},
|
|
{
|
|
"epoch": 7.085106382978723,
|
|
"grad_norm": 1.0895054515077724,
|
|
"learning_rate": 1.0589192095755172e-06,
|
|
"loss": 0.0217,
|
|
"step": 333
|
|
},
|
|
{
|
|
"epoch": 7.1063829787234045,
|
|
"grad_norm": 1.0138202696131422,
|
|
"learning_rate": 1.0446925746067768e-06,
|
|
"loss": 0.0189,
|
|
"step": 334
|
|
},
|
|
{
|
|
"epoch": 7.127659574468085,
|
|
"grad_norm": 0.931336055385696,
|
|
"learning_rate": 1.0305368692688175e-06,
|
|
"loss": 0.0209,
|
|
"step": 335
|
|
},
|
|
{
|
|
"epoch": 7.148936170212766,
|
|
"grad_norm": 1.214374992403074,
|
|
"learning_rate": 1.0164527834907468e-06,
|
|
"loss": 0.0218,
|
|
"step": 336
|
|
},
|
|
{
|
|
"epoch": 7.170212765957447,
|
|
"grad_norm": 1.5245924187989146,
|
|
"learning_rate": 1.0024410037110358e-06,
|
|
"loss": 0.022,
|
|
"step": 337
|
|
},
|
|
{
|
|
"epoch": 7.191489361702128,
|
|
"grad_norm": 1.0820503643925237,
|
|
"learning_rate": 9.88502212844063e-07,
|
|
"loss": 0.0171,
|
|
"step": 338
|
|
},
|
|
{
|
|
"epoch": 7.212765957446808,
|
|
"grad_norm": 0.8009422999145568,
|
|
"learning_rate": 9.746370902468311e-07,
|
|
"loss": 0.0179,
|
|
"step": 339
|
|
},
|
|
{
|
|
"epoch": 7.23404255319149,
|
|
"grad_norm": 1.245836157318975,
|
|
"learning_rate": 9.608463116858544e-07,
|
|
"loss": 0.0196,
|
|
"step": 340
|
|
},
|
|
{
|
|
"epoch": 7.25531914893617,
|
|
"grad_norm": 0.9189114786673717,
|
|
"learning_rate": 9.471305493042243e-07,
|
|
"loss": 0.0206,
|
|
"step": 341
|
|
},
|
|
{
|
|
"epoch": 7.276595744680851,
|
|
"grad_norm": 1.2781854766390135,
|
|
"learning_rate": 9.334904715888496e-07,
|
|
"loss": 0.0164,
|
|
"step": 342
|
|
},
|
|
{
|
|
"epoch": 7.297872340425532,
|
|
"grad_norm": 1.5271374445001962,
|
|
"learning_rate": 9.199267433378728e-07,
|
|
"loss": 0.02,
|
|
"step": 343
|
|
},
|
|
{
|
|
"epoch": 7.319148936170213,
|
|
"grad_norm": 0.946456072976271,
|
|
"learning_rate": 9.064400256282757e-07,
|
|
"loss": 0.0214,
|
|
"step": 344
|
|
},
|
|
{
|
|
"epoch": 7.340425531914893,
|
|
"grad_norm": 1.270831674714735,
|
|
"learning_rate": 8.930309757836517e-07,
|
|
"loss": 0.0216,
|
|
"step": 345
|
|
},
|
|
{
|
|
"epoch": 7.361702127659575,
|
|
"grad_norm": 0.9172465899189646,
|
|
"learning_rate": 8.797002473421729e-07,
|
|
"loss": 0.0172,
|
|
"step": 346
|
|
},
|
|
{
|
|
"epoch": 7.382978723404255,
|
|
"grad_norm": 0.9762896610645739,
|
|
"learning_rate": 8.664484900247363e-07,
|
|
"loss": 0.0196,
|
|
"step": 347
|
|
},
|
|
{
|
|
"epoch": 7.404255319148936,
|
|
"grad_norm": 1.2366903890873284,
|
|
"learning_rate": 8.532763497032987e-07,
|
|
"loss": 0.0192,
|
|
"step": 348
|
|
},
|
|
{
|
|
"epoch": 7.425531914893617,
|
|
"grad_norm": 0.8921247183956026,
|
|
"learning_rate": 8.40184468369396e-07,
|
|
"loss": 0.0138,
|
|
"step": 349
|
|
},
|
|
{
|
|
"epoch": 7.446808510638298,
|
|
"grad_norm": 1.371683786414812,
|
|
"learning_rate": 8.271734841028553e-07,
|
|
"loss": 0.0257,
|
|
"step": 350
|
|
},
|
|
{
|
|
"epoch": 7.468085106382979,
|
|
"grad_norm": 1.6706279670454172,
|
|
"learning_rate": 8.142440310406923e-07,
|
|
"loss": 0.0221,
|
|
"step": 351
|
|
},
|
|
{
|
|
"epoch": 7.48936170212766,
|
|
"grad_norm": 1.0048836984452958,
|
|
"learning_rate": 8.013967393462094e-07,
|
|
"loss": 0.0179,
|
|
"step": 352
|
|
},
|
|
{
|
|
"epoch": 7.51063829787234,
|
|
"grad_norm": 1.028897861820127,
|
|
"learning_rate": 7.886322351782782e-07,
|
|
"loss": 0.0173,
|
|
"step": 353
|
|
},
|
|
{
|
|
"epoch": 7.531914893617021,
|
|
"grad_norm": 0.9414634888570663,
|
|
"learning_rate": 7.759511406608255e-07,
|
|
"loss": 0.0194,
|
|
"step": 354
|
|
},
|
|
{
|
|
"epoch": 7.553191489361702,
|
|
"grad_norm": 1.3421193366601667,
|
|
"learning_rate": 7.633540738525066e-07,
|
|
"loss": 0.0155,
|
|
"step": 355
|
|
},
|
|
{
|
|
"epoch": 7.574468085106383,
|
|
"grad_norm": 1.3991724028833572,
|
|
"learning_rate": 7.508416487165862e-07,
|
|
"loss": 0.0242,
|
|
"step": 356
|
|
},
|
|
{
|
|
"epoch": 7.595744680851064,
|
|
"grad_norm": 3.4908651855373067,
|
|
"learning_rate": 7.384144750910133e-07,
|
|
"loss": 0.0201,
|
|
"step": 357
|
|
},
|
|
{
|
|
"epoch": 7.617021276595745,
|
|
"grad_norm": 1.1125708607058544,
|
|
"learning_rate": 7.260731586586983e-07,
|
|
"loss": 0.0181,
|
|
"step": 358
|
|
},
|
|
{
|
|
"epoch": 7.638297872340425,
|
|
"grad_norm": 2.003422428134198,
|
|
"learning_rate": 7.138183009179922e-07,
|
|
"loss": 0.0244,
|
|
"step": 359
|
|
},
|
|
{
|
|
"epoch": 7.659574468085106,
|
|
"grad_norm": 0.8666069576765489,
|
|
"learning_rate": 7.016504991533727e-07,
|
|
"loss": 0.0231,
|
|
"step": 360
|
|
},
|
|
{
|
|
"epoch": 7.680851063829787,
|
|
"grad_norm": 0.8225857804515108,
|
|
"learning_rate": 6.895703464063319e-07,
|
|
"loss": 0.0127,
|
|
"step": 361
|
|
},
|
|
{
|
|
"epoch": 7.702127659574468,
|
|
"grad_norm": 1.2220101527746525,
|
|
"learning_rate": 6.775784314464717e-07,
|
|
"loss": 0.0261,
|
|
"step": 362
|
|
},
|
|
{
|
|
"epoch": 7.723404255319149,
|
|
"grad_norm": 1.291947063153676,
|
|
"learning_rate": 6.656753387428089e-07,
|
|
"loss": 0.024,
|
|
"step": 363
|
|
},
|
|
{
|
|
"epoch": 7.74468085106383,
|
|
"grad_norm": 1.288638995380391,
|
|
"learning_rate": 6.538616484352902e-07,
|
|
"loss": 0.016,
|
|
"step": 364
|
|
},
|
|
{
|
|
"epoch": 7.76595744680851,
|
|
"grad_norm": 1.3674238710009354,
|
|
"learning_rate": 6.421379363065142e-07,
|
|
"loss": 0.0253,
|
|
"step": 365
|
|
},
|
|
{
|
|
"epoch": 7.787234042553192,
|
|
"grad_norm": 0.877515831783811,
|
|
"learning_rate": 6.305047737536707e-07,
|
|
"loss": 0.0162,
|
|
"step": 366
|
|
},
|
|
{
|
|
"epoch": 7.808510638297872,
|
|
"grad_norm": 1.1630712659406794,
|
|
"learning_rate": 6.189627277606894e-07,
|
|
"loss": 0.0227,
|
|
"step": 367
|
|
},
|
|
{
|
|
"epoch": 7.829787234042553,
|
|
"grad_norm": 1.2379019290179472,
|
|
"learning_rate": 6.075123608706093e-07,
|
|
"loss": 0.0179,
|
|
"step": 368
|
|
},
|
|
{
|
|
"epoch": 7.851063829787234,
|
|
"grad_norm": 0.9761836229980341,
|
|
"learning_rate": 5.961542311581586e-07,
|
|
"loss": 0.017,
|
|
"step": 369
|
|
},
|
|
{
|
|
"epoch": 7.872340425531915,
|
|
"grad_norm": 1.2905446310752584,
|
|
"learning_rate": 5.848888922025553e-07,
|
|
"loss": 0.0237,
|
|
"step": 370
|
|
},
|
|
{
|
|
"epoch": 7.8936170212765955,
|
|
"grad_norm": 1.0599111161417107,
|
|
"learning_rate": 5.737168930605272e-07,
|
|
"loss": 0.0193,
|
|
"step": 371
|
|
},
|
|
{
|
|
"epoch": 7.914893617021277,
|
|
"grad_norm": 0.9002316779352258,
|
|
"learning_rate": 5.626387782395512e-07,
|
|
"loss": 0.0219,
|
|
"step": 372
|
|
},
|
|
{
|
|
"epoch": 7.9361702127659575,
|
|
"grad_norm": 1.0871181403304409,
|
|
"learning_rate": 5.516550876713142e-07,
|
|
"loss": 0.0202,
|
|
"step": 373
|
|
},
|
|
{
|
|
"epoch": 7.957446808510638,
|
|
"grad_norm": 0.7739867037107642,
|
|
"learning_rate": 5.407663566854008e-07,
|
|
"loss": 0.0141,
|
|
"step": 374
|
|
},
|
|
{
|
|
"epoch": 7.9787234042553195,
|
|
"grad_norm": 1.1646231525599027,
|
|
"learning_rate": 5.299731159831953e-07,
|
|
"loss": 0.0199,
|
|
"step": 375
|
|
},
|
|
{
|
|
"epoch": 8.0,
|
|
"grad_norm": 0.8486174354933951,
|
|
"learning_rate": 5.192758916120236e-07,
|
|
"loss": 0.0183,
|
|
"step": 376
|
|
},
|
|
{
|
|
"epoch": 8.02127659574468,
|
|
"grad_norm": 0.5857908765113855,
|
|
"learning_rate": 5.086752049395094e-07,
|
|
"loss": 0.0154,
|
|
"step": 377
|
|
},
|
|
{
|
|
"epoch": 8.042553191489361,
|
|
"grad_norm": 0.820073038781464,
|
|
"learning_rate": 4.981715726281666e-07,
|
|
"loss": 0.0162,
|
|
"step": 378
|
|
},
|
|
{
|
|
"epoch": 8.063829787234043,
|
|
"grad_norm": 0.5353750025209411,
|
|
"learning_rate": 4.87765506610215e-07,
|
|
"loss": 0.0117,
|
|
"step": 379
|
|
},
|
|
{
|
|
"epoch": 8.085106382978724,
|
|
"grad_norm": 0.6641723492880932,
|
|
"learning_rate": 4.774575140626317e-07,
|
|
"loss": 0.0155,
|
|
"step": 380
|
|
},
|
|
{
|
|
"epoch": 8.106382978723405,
|
|
"grad_norm": 1.5478055693073858,
|
|
"learning_rate": 4.672480973824312e-07,
|
|
"loss": 0.0138,
|
|
"step": 381
|
|
},
|
|
{
|
|
"epoch": 8.127659574468085,
|
|
"grad_norm": 0.5935452327954517,
|
|
"learning_rate": 4.5713775416217884e-07,
|
|
"loss": 0.0128,
|
|
"step": 382
|
|
},
|
|
{
|
|
"epoch": 8.148936170212766,
|
|
"grad_norm": 0.45835158236977624,
|
|
"learning_rate": 4.4712697716573994e-07,
|
|
"loss": 0.014,
|
|
"step": 383
|
|
},
|
|
{
|
|
"epoch": 8.170212765957446,
|
|
"grad_norm": 0.495302571001091,
|
|
"learning_rate": 4.372162543042624e-07,
|
|
"loss": 0.0135,
|
|
"step": 384
|
|
},
|
|
{
|
|
"epoch": 8.191489361702128,
|
|
"grad_norm": 0.4653447676593763,
|
|
"learning_rate": 4.27406068612396e-07,
|
|
"loss": 0.0113,
|
|
"step": 385
|
|
},
|
|
{
|
|
"epoch": 8.212765957446809,
|
|
"grad_norm": 0.5764443378886089,
|
|
"learning_rate": 4.1769689822475147e-07,
|
|
"loss": 0.0149,
|
|
"step": 386
|
|
},
|
|
{
|
|
"epoch": 8.23404255319149,
|
|
"grad_norm": 0.34507606047404416,
|
|
"learning_rate": 4.0808921635259595e-07,
|
|
"loss": 0.0096,
|
|
"step": 387
|
|
},
|
|
{
|
|
"epoch": 8.25531914893617,
|
|
"grad_norm": 0.6316918421422608,
|
|
"learning_rate": 3.9858349126078945e-07,
|
|
"loss": 0.0149,
|
|
"step": 388
|
|
},
|
|
{
|
|
"epoch": 8.27659574468085,
|
|
"grad_norm": 1.13945433310328,
|
|
"learning_rate": 3.891801862449629e-07,
|
|
"loss": 0.0171,
|
|
"step": 389
|
|
},
|
|
{
|
|
"epoch": 8.297872340425531,
|
|
"grad_norm": 1.6671178791582812,
|
|
"learning_rate": 3.798797596089351e-07,
|
|
"loss": 0.0162,
|
|
"step": 390
|
|
},
|
|
{
|
|
"epoch": 8.319148936170214,
|
|
"grad_norm": 0.6242884167930942,
|
|
"learning_rate": 3.7068266464238085e-07,
|
|
"loss": 0.0171,
|
|
"step": 391
|
|
},
|
|
{
|
|
"epoch": 8.340425531914894,
|
|
"grad_norm": 0.6157189166745389,
|
|
"learning_rate": 3.615893495987335e-07,
|
|
"loss": 0.0083,
|
|
"step": 392
|
|
},
|
|
{
|
|
"epoch": 8.361702127659575,
|
|
"grad_norm": 1.187375482456905,
|
|
"learning_rate": 3.5260025767333894e-07,
|
|
"loss": 0.0133,
|
|
"step": 393
|
|
},
|
|
{
|
|
"epoch": 8.382978723404255,
|
|
"grad_norm": 0.9198827823698973,
|
|
"learning_rate": 3.4371582698185636e-07,
|
|
"loss": 0.0168,
|
|
"step": 394
|
|
},
|
|
{
|
|
"epoch": 8.404255319148936,
|
|
"grad_norm": 0.9127337383828107,
|
|
"learning_rate": 3.3493649053890325e-07,
|
|
"loss": 0.0167,
|
|
"step": 395
|
|
},
|
|
{
|
|
"epoch": 8.425531914893616,
|
|
"grad_norm": 0.4716697636311403,
|
|
"learning_rate": 3.262626762369525e-07,
|
|
"loss": 0.0134,
|
|
"step": 396
|
|
},
|
|
{
|
|
"epoch": 8.446808510638299,
|
|
"grad_norm": 0.5430698963916597,
|
|
"learning_rate": 3.176948068254762e-07,
|
|
"loss": 0.0115,
|
|
"step": 397
|
|
},
|
|
{
|
|
"epoch": 8.46808510638298,
|
|
"grad_norm": 0.5589379122310724,
|
|
"learning_rate": 3.092332998903416e-07,
|
|
"loss": 0.0134,
|
|
"step": 398
|
|
},
|
|
{
|
|
"epoch": 8.48936170212766,
|
|
"grad_norm": 1.0297905449956246,
|
|
"learning_rate": 3.0087856783345916e-07,
|
|
"loss": 0.0165,
|
|
"step": 399
|
|
},
|
|
{
|
|
"epoch": 8.51063829787234,
|
|
"grad_norm": 0.5586126342363132,
|
|
"learning_rate": 2.9263101785268253e-07,
|
|
"loss": 0.0129,
|
|
"step": 400
|
|
},
|
|
{
|
|
"epoch": 8.51063829787234,
|
|
"eval_loss": 1.6673990488052368,
|
|
"eval_runtime": 1.6235,
|
|
"eval_samples_per_second": 14.167,
|
|
"eval_steps_per_second": 0.616,
|
|
"step": 400
|
|
},
|
|
{
|
|
"epoch": 8.53191489361702,
|
|
"grad_norm": 0.5428039786664497,
|
|
"learning_rate": 2.844910519219632e-07,
|
|
"loss": 0.0128,
|
|
"step": 401
|
|
},
|
|
{
|
|
"epoch": 8.553191489361701,
|
|
"grad_norm": 0.633307262716061,
|
|
"learning_rate": 2.764590667717562e-07,
|
|
"loss": 0.0161,
|
|
"step": 402
|
|
},
|
|
{
|
|
"epoch": 8.574468085106384,
|
|
"grad_norm": 0.5251963830034382,
|
|
"learning_rate": 2.6853545386968607e-07,
|
|
"loss": 0.0115,
|
|
"step": 403
|
|
},
|
|
{
|
|
"epoch": 8.595744680851064,
|
|
"grad_norm": 0.48124935494240734,
|
|
"learning_rate": 2.6072059940146775e-07,
|
|
"loss": 0.0097,
|
|
"step": 404
|
|
},
|
|
{
|
|
"epoch": 8.617021276595745,
|
|
"grad_norm": 0.5546850016440125,
|
|
"learning_rate": 2.53014884252083e-07,
|
|
"loss": 0.0095,
|
|
"step": 405
|
|
},
|
|
{
|
|
"epoch": 8.638297872340425,
|
|
"grad_norm": 0.6870597575288601,
|
|
"learning_rate": 2.454186839872158e-07,
|
|
"loss": 0.0161,
|
|
"step": 406
|
|
},
|
|
{
|
|
"epoch": 8.659574468085106,
|
|
"grad_norm": 0.5075138452670342,
|
|
"learning_rate": 2.3793236883495164e-07,
|
|
"loss": 0.0137,
|
|
"step": 407
|
|
},
|
|
{
|
|
"epoch": 8.680851063829786,
|
|
"grad_norm": 0.7184250889463067,
|
|
"learning_rate": 2.3055630366772857e-07,
|
|
"loss": 0.0124,
|
|
"step": 408
|
|
},
|
|
{
|
|
"epoch": 8.702127659574469,
|
|
"grad_norm": 0.6095068862526991,
|
|
"learning_rate": 2.2329084798455747e-07,
|
|
"loss": 0.0145,
|
|
"step": 409
|
|
},
|
|
{
|
|
"epoch": 8.72340425531915,
|
|
"grad_norm": 0.5513423746444825,
|
|
"learning_rate": 2.1613635589349756e-07,
|
|
"loss": 0.0155,
|
|
"step": 410
|
|
},
|
|
{
|
|
"epoch": 8.74468085106383,
|
|
"grad_norm": 0.5875898983901224,
|
|
"learning_rate": 2.0909317609440093e-07,
|
|
"loss": 0.0146,
|
|
"step": 411
|
|
},
|
|
{
|
|
"epoch": 8.76595744680851,
|
|
"grad_norm": 0.5750560265603247,
|
|
"learning_rate": 2.0216165186191406e-07,
|
|
"loss": 0.0131,
|
|
"step": 412
|
|
},
|
|
{
|
|
"epoch": 8.787234042553191,
|
|
"grad_norm": 0.6746724848758707,
|
|
"learning_rate": 1.95342121028749e-07,
|
|
"loss": 0.0118,
|
|
"step": 413
|
|
},
|
|
{
|
|
"epoch": 8.808510638297872,
|
|
"grad_norm": 0.6972301523231518,
|
|
"learning_rate": 1.8863491596921745e-07,
|
|
"loss": 0.0143,
|
|
"step": 414
|
|
},
|
|
{
|
|
"epoch": 8.829787234042554,
|
|
"grad_norm": 0.7525328974002997,
|
|
"learning_rate": 1.8204036358303173e-07,
|
|
"loss": 0.0144,
|
|
"step": 415
|
|
},
|
|
{
|
|
"epoch": 8.851063829787234,
|
|
"grad_norm": 0.5087036176440635,
|
|
"learning_rate": 1.7555878527937164e-07,
|
|
"loss": 0.0117,
|
|
"step": 416
|
|
},
|
|
{
|
|
"epoch": 8.872340425531915,
|
|
"grad_norm": 0.6061494141417032,
|
|
"learning_rate": 1.6919049696121957e-07,
|
|
"loss": 0.0168,
|
|
"step": 417
|
|
},
|
|
{
|
|
"epoch": 8.893617021276595,
|
|
"grad_norm": 0.6674633587572225,
|
|
"learning_rate": 1.629358090099639e-07,
|
|
"loss": 0.0154,
|
|
"step": 418
|
|
},
|
|
{
|
|
"epoch": 8.914893617021276,
|
|
"grad_norm": 0.6381527241007888,
|
|
"learning_rate": 1.567950262702714e-07,
|
|
"loss": 0.0118,
|
|
"step": 419
|
|
},
|
|
{
|
|
"epoch": 8.936170212765958,
|
|
"grad_norm": 0.5652826587480684,
|
|
"learning_rate": 1.507684480352292e-07,
|
|
"loss": 0.0142,
|
|
"step": 420
|
|
},
|
|
{
|
|
"epoch": 8.957446808510639,
|
|
"grad_norm": 0.5390136780396995,
|
|
"learning_rate": 1.4485636803175828e-07,
|
|
"loss": 0.0153,
|
|
"step": 421
|
|
},
|
|
{
|
|
"epoch": 8.97872340425532,
|
|
"grad_norm": 0.4433110433252025,
|
|
"learning_rate": 1.3905907440629752e-07,
|
|
"loss": 0.0128,
|
|
"step": 422
|
|
},
|
|
{
|
|
"epoch": 9.0,
|
|
"grad_norm": 0.3937254962258607,
|
|
"learning_rate": 1.3337684971075932e-07,
|
|
"loss": 0.0141,
|
|
"step": 423
|
|
},
|
|
{
|
|
"epoch": 9.02127659574468,
|
|
"grad_norm": 0.42872899914768886,
|
|
"learning_rate": 1.278099708887587e-07,
|
|
"loss": 0.0139,
|
|
"step": 424
|
|
},
|
|
{
|
|
"epoch": 9.042553191489361,
|
|
"grad_norm": 0.43176685764328526,
|
|
"learning_rate": 1.223587092621162e-07,
|
|
"loss": 0.01,
|
|
"step": 425
|
|
},
|
|
{
|
|
"epoch": 9.063829787234043,
|
|
"grad_norm": 0.3743485104670554,
|
|
"learning_rate": 1.1702333051763271e-07,
|
|
"loss": 0.0082,
|
|
"step": 426
|
|
},
|
|
{
|
|
"epoch": 9.085106382978724,
|
|
"grad_norm": 0.39987032601374245,
|
|
"learning_rate": 1.1180409469414094e-07,
|
|
"loss": 0.0125,
|
|
"step": 427
|
|
},
|
|
{
|
|
"epoch": 9.106382978723405,
|
|
"grad_norm": 0.39568260783662074,
|
|
"learning_rate": 1.067012561698319e-07,
|
|
"loss": 0.0099,
|
|
"step": 428
|
|
},
|
|
{
|
|
"epoch": 9.127659574468085,
|
|
"grad_norm": 0.4002164747878374,
|
|
"learning_rate": 1.0171506364985622e-07,
|
|
"loss": 0.0141,
|
|
"step": 429
|
|
},
|
|
{
|
|
"epoch": 9.148936170212766,
|
|
"grad_norm": 0.3140844748261339,
|
|
"learning_rate": 9.684576015420277e-08,
|
|
"loss": 0.0083,
|
|
"step": 430
|
|
},
|
|
{
|
|
"epoch": 9.170212765957446,
|
|
"grad_norm": 3.0839571989766026,
|
|
"learning_rate": 9.209358300585474e-08,
|
|
"loss": 0.0107,
|
|
"step": 431
|
|
},
|
|
{
|
|
"epoch": 9.191489361702128,
|
|
"grad_norm": 0.43386426711502496,
|
|
"learning_rate": 8.745876381922147e-08,
|
|
"loss": 0.0114,
|
|
"step": 432
|
|
},
|
|
{
|
|
"epoch": 9.212765957446809,
|
|
"grad_norm": 0.49575077577441945,
|
|
"learning_rate": 8.294152848885156e-08,
|
|
"loss": 0.0122,
|
|
"step": 433
|
|
},
|
|
{
|
|
"epoch": 9.23404255319149,
|
|
"grad_norm": 0.42605421543375827,
|
|
"learning_rate": 7.854209717842231e-08,
|
|
"loss": 0.0123,
|
|
"step": 434
|
|
},
|
|
{
|
|
"epoch": 9.25531914893617,
|
|
"grad_norm": 0.4159085198061852,
|
|
"learning_rate": 7.426068431000883e-08,
|
|
"loss": 0.011,
|
|
"step": 435
|
|
},
|
|
{
|
|
"epoch": 9.27659574468085,
|
|
"grad_norm": 0.4297510739797917,
|
|
"learning_rate": 7.009749855363457e-08,
|
|
"loss": 0.0096,
|
|
"step": 436
|
|
},
|
|
{
|
|
"epoch": 9.297872340425531,
|
|
"grad_norm": 0.39770284394181865,
|
|
"learning_rate": 6.605274281709929e-08,
|
|
"loss": 0.0108,
|
|
"step": 437
|
|
},
|
|
{
|
|
"epoch": 9.319148936170214,
|
|
"grad_norm": 0.3799358013058963,
|
|
"learning_rate": 6.212661423609184e-08,
|
|
"loss": 0.0135,
|
|
"step": 438
|
|
},
|
|
{
|
|
"epoch": 9.340425531914894,
|
|
"grad_norm": 0.40685286788703806,
|
|
"learning_rate": 5.83193041645802e-08,
|
|
"loss": 0.0105,
|
|
"step": 439
|
|
},
|
|
{
|
|
"epoch": 9.361702127659575,
|
|
"grad_norm": 0.5276489442708509,
|
|
"learning_rate": 5.463099816548578e-08,
|
|
"loss": 0.0159,
|
|
"step": 440
|
|
},
|
|
{
|
|
"epoch": 9.382978723404255,
|
|
"grad_norm": 0.3398816914667588,
|
|
"learning_rate": 5.106187600163987e-08,
|
|
"loss": 0.01,
|
|
"step": 441
|
|
},
|
|
{
|
|
"epoch": 9.404255319148936,
|
|
"grad_norm": 0.43060715328426347,
|
|
"learning_rate": 4.761211162702117e-08,
|
|
"loss": 0.0109,
|
|
"step": 442
|
|
},
|
|
{
|
|
"epoch": 9.425531914893616,
|
|
"grad_norm": 0.6155281563788073,
|
|
"learning_rate": 4.428187317827848e-08,
|
|
"loss": 0.0107,
|
|
"step": 443
|
|
},
|
|
{
|
|
"epoch": 9.446808510638299,
|
|
"grad_norm": 0.4920360355401412,
|
|
"learning_rate": 4.1071322966535487e-08,
|
|
"loss": 0.0139,
|
|
"step": 444
|
|
},
|
|
{
|
|
"epoch": 9.46808510638298,
|
|
"grad_norm": 0.44825099453614115,
|
|
"learning_rate": 3.798061746947995e-08,
|
|
"loss": 0.0113,
|
|
"step": 445
|
|
},
|
|
{
|
|
"epoch": 9.48936170212766,
|
|
"grad_norm": 0.48315881463136157,
|
|
"learning_rate": 3.5009907323737826e-08,
|
|
"loss": 0.0131,
|
|
"step": 446
|
|
},
|
|
{
|
|
"epoch": 9.51063829787234,
|
|
"grad_norm": 0.3897061025707457,
|
|
"learning_rate": 3.2159337317530234e-08,
|
|
"loss": 0.0131,
|
|
"step": 447
|
|
},
|
|
{
|
|
"epoch": 9.53191489361702,
|
|
"grad_norm": 0.3845441440160021,
|
|
"learning_rate": 2.9429046383618042e-08,
|
|
"loss": 0.0094,
|
|
"step": 448
|
|
},
|
|
{
|
|
"epoch": 9.553191489361701,
|
|
"grad_norm": 0.4018776472448016,
|
|
"learning_rate": 2.681916759252917e-08,
|
|
"loss": 0.0117,
|
|
"step": 449
|
|
},
|
|
{
|
|
"epoch": 9.574468085106384,
|
|
"grad_norm": 0.3789922116838833,
|
|
"learning_rate": 2.4329828146074096e-08,
|
|
"loss": 0.0099,
|
|
"step": 450
|
|
},
|
|
{
|
|
"epoch": 9.595744680851064,
|
|
"grad_norm": 0.47356186931300714,
|
|
"learning_rate": 2.1961149371145795e-08,
|
|
"loss": 0.012,
|
|
"step": 451
|
|
},
|
|
{
|
|
"epoch": 9.617021276595745,
|
|
"grad_norm": 0.5175691416116797,
|
|
"learning_rate": 1.9713246713805588e-08,
|
|
"loss": 0.0137,
|
|
"step": 452
|
|
},
|
|
{
|
|
"epoch": 9.638297872340425,
|
|
"grad_norm": 0.35550871415754165,
|
|
"learning_rate": 1.7586229733657646e-08,
|
|
"loss": 0.0126,
|
|
"step": 453
|
|
},
|
|
{
|
|
"epoch": 9.659574468085106,
|
|
"grad_norm": 0.4521940543192018,
|
|
"learning_rate": 1.5580202098509078e-08,
|
|
"loss": 0.0123,
|
|
"step": 454
|
|
},
|
|
{
|
|
"epoch": 9.680851063829786,
|
|
"grad_norm": 0.3645510186204655,
|
|
"learning_rate": 1.3695261579316776e-08,
|
|
"loss": 0.0099,
|
|
"step": 455
|
|
},
|
|
{
|
|
"epoch": 9.702127659574469,
|
|
"grad_norm": 0.5052016943572653,
|
|
"learning_rate": 1.193150004542204e-08,
|
|
"loss": 0.0108,
|
|
"step": 456
|
|
},
|
|
{
|
|
"epoch": 9.72340425531915,
|
|
"grad_norm": 0.4591420783076054,
|
|
"learning_rate": 1.0289003460074165e-08,
|
|
"loss": 0.0136,
|
|
"step": 457
|
|
},
|
|
{
|
|
"epoch": 9.74468085106383,
|
|
"grad_norm": 0.3979451604394145,
|
|
"learning_rate": 8.767851876239075e-09,
|
|
"loss": 0.0124,
|
|
"step": 458
|
|
},
|
|
{
|
|
"epoch": 9.76595744680851,
|
|
"grad_norm": 0.4508311448933257,
|
|
"learning_rate": 7.368119432699383e-09,
|
|
"loss": 0.0117,
|
|
"step": 459
|
|
},
|
|
{
|
|
"epoch": 9.787234042553191,
|
|
"grad_norm": 0.48162461236003445,
|
|
"learning_rate": 6.089874350439507e-09,
|
|
"loss": 0.0131,
|
|
"step": 460
|
|
},
|
|
{
|
|
"epoch": 9.808510638297872,
|
|
"grad_norm": 0.3727617520322409,
|
|
"learning_rate": 4.933178929321103e-09,
|
|
"loss": 0.0092,
|
|
"step": 461
|
|
},
|
|
{
|
|
"epoch": 9.829787234042554,
|
|
"grad_norm": 0.37748204227545856,
|
|
"learning_rate": 3.8980895450474455e-09,
|
|
"loss": 0.0108,
|
|
"step": 462
|
|
},
|
|
{
|
|
"epoch": 9.851063829787234,
|
|
"grad_norm": 0.30954369790579545,
|
|
"learning_rate": 2.984656646415063e-09,
|
|
"loss": 0.0082,
|
|
"step": 463
|
|
},
|
|
{
|
|
"epoch": 9.872340425531915,
|
|
"grad_norm": 0.441565820890093,
|
|
"learning_rate": 2.192924752854042e-09,
|
|
"loss": 0.014,
|
|
"step": 464
|
|
},
|
|
{
|
|
"epoch": 9.893617021276595,
|
|
"grad_norm": 0.42642106144925446,
|
|
"learning_rate": 1.5229324522605949e-09,
|
|
"loss": 0.0086,
|
|
"step": 465
|
|
},
|
|
{
|
|
"epoch": 9.914893617021276,
|
|
"grad_norm": 0.3393626098420996,
|
|
"learning_rate": 9.747123991141193e-10,
|
|
"loss": 0.008,
|
|
"step": 466
|
|
},
|
|
{
|
|
"epoch": 9.936170212765958,
|
|
"grad_norm": 0.4441469010998739,
|
|
"learning_rate": 5.48291312886251e-10,
|
|
"loss": 0.0135,
|
|
"step": 467
|
|
},
|
|
{
|
|
"epoch": 9.957446808510639,
|
|
"grad_norm": 0.44011051414564956,
|
|
"learning_rate": 2.43689976739403e-10,
|
|
"loss": 0.0117,
|
|
"step": 468
|
|
},
|
|
{
|
|
"epoch": 9.97872340425532,
|
|
"grad_norm": 0.47111655728750834,
|
|
"learning_rate": 6.092323651313293e-11,
|
|
"loss": 0.0104,
|
|
"step": 469
|
|
},
|
|
{
|
|
"epoch": 10.0,
|
|
"grad_norm": 0.5133248194623898,
|
|
"learning_rate": 0.0,
|
|
"loss": 0.0161,
|
|
"step": 470
|
|
}
|
|
],
|
|
"logging_steps": 1.0,
|
|
"max_steps": 470,
|
|
"num_input_tokens_seen": 0,
|
|
"num_train_epochs": 10,
|
|
"save_steps": 1000,
|
|
"stateful_callbacks": {
|
|
"TrainerControl": {
|
|
"args": {
|
|
"should_epoch_stop": false,
|
|
"should_evaluate": false,
|
|
"should_log": false,
|
|
"should_save": true,
|
|
"should_training_stop": true
|
|
},
|
|
"attributes": {}
|
|
}
|
|
},
|
|
"total_flos": 32811979898880.0,
|
|
"train_batch_size": 6,
|
|
"trial_name": null,
|
|
"trial_params": null
|
|
}
|