Files
ModelHub XC 665c0870aa 初始化项目,由ModelHub XC社区提供模型
Model: cjiao/OpenThoughts3-random-groups-openthinker3-1.5B-checkpoint-375-80
Source: Original Platform
2026-04-28 10:59:02 +08:00

1154 lines
26 KiB
JSON

{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.0,
"eval_steps": 500,
"global_step": 160,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0125,
"grad_norm": 2.2997210025787354,
"learning_rate": 0.00015998457923856519,
"loss": 1.2401,
"step": 1
},
{
"epoch": 0.025,
"grad_norm": 24.28518295288086,
"learning_rate": 0.00015993832289925785,
"loss": 4.0648,
"step": 2
},
{
"epoch": 0.0375,
"grad_norm": 12.727800369262695,
"learning_rate": 0.0001598612488147773,
"loss": 2.7044,
"step": 3
},
{
"epoch": 0.05,
"grad_norm": 4.934032440185547,
"learning_rate": 0.00015975338669865026,
"loss": 1.994,
"step": 4
},
{
"epoch": 0.0625,
"grad_norm": 8.1886625289917,
"learning_rate": 0.00015961477813377576,
"loss": 2.1426,
"step": 5
},
{
"epoch": 0.075,
"grad_norm": 2.910872220993042,
"learning_rate": 0.00015944547655639412,
"loss": 1.7254,
"step": 6
},
{
"epoch": 0.0875,
"grad_norm": 1.565290093421936,
"learning_rate": 0.00015924554723548617,
"loss": 1.5187,
"step": 7
},
{
"epoch": 0.1,
"grad_norm": 2.9080827236175537,
"learning_rate": 0.00015901506724761103,
"loss": 1.5405,
"step": 8
},
{
"epoch": 0.1125,
"grad_norm": 1.875126838684082,
"learning_rate": 0.00015875412544719134,
"loss": 1.4493,
"step": 9
},
{
"epoch": 0.125,
"grad_norm": 1.2589935064315796,
"learning_rate": 0.00015846282243225845,
"loss": 1.4103,
"step": 10
},
{
"epoch": 0.1375,
"grad_norm": 1.2925529479980469,
"learning_rate": 0.0001581412705056698,
"loss": 1.3507,
"step": 11
},
{
"epoch": 0.15,
"grad_norm": 1.4467802047729492,
"learning_rate": 0.00015778959363181415,
"loss": 1.3,
"step": 12
},
{
"epoch": 0.1625,
"grad_norm": 1.267639398574829,
"learning_rate": 0.0001574079273888208,
"loss": 1.2974,
"step": 13
},
{
"epoch": 0.175,
"grad_norm": 1.0911085605621338,
"learning_rate": 0.00015699641891629178,
"loss": 1.2635,
"step": 14
},
{
"epoch": 0.1875,
"grad_norm": 0.9065354466438293,
"learning_rate": 0.00015655522685857672,
"loss": 1.2119,
"step": 15
},
{
"epoch": 0.2,
"grad_norm": 0.7415559887886047,
"learning_rate": 0.0001560845213036123,
"loss": 1.2337,
"step": 16
},
{
"epoch": 0.2125,
"grad_norm": 0.7553166151046753,
"learning_rate": 0.00015558448371735025,
"loss": 1.1884,
"step": 17
},
{
"epoch": 0.225,
"grad_norm": 0.5407947301864624,
"learning_rate": 0.00015505530687379875,
"loss": 1.17,
"step": 18
},
{
"epoch": 0.2375,
"grad_norm": 0.5162355899810791,
"learning_rate": 0.00015449719478070428,
"loss": 1.1879,
"step": 19
},
{
"epoch": 0.25,
"grad_norm": 0.5688554644584656,
"learning_rate": 0.00015391036260090294,
"loss": 1.1767,
"step": 20
},
{
"epoch": 0.2625,
"grad_norm": 0.48555564880371094,
"learning_rate": 0.0001532950365693709,
"loss": 1.1726,
"step": 21
},
{
"epoch": 0.275,
"grad_norm": 0.4502723515033722,
"learning_rate": 0.00015265145390600652,
"loss": 1.163,
"step": 22
},
{
"epoch": 0.2875,
"grad_norm": 0.3590157926082611,
"learning_rate": 0.00015197986272417774,
"loss": 1.1839,
"step": 23
},
{
"epoch": 0.3,
"grad_norm": 0.38364410400390625,
"learning_rate": 0.00015128052193506944,
"loss": 1.1642,
"step": 24
},
{
"epoch": 0.3125,
"grad_norm": 0.36856546998023987,
"learning_rate": 0.0001505537011478684,
"loss": 1.1495,
"step": 25
},
{
"epoch": 0.325,
"grad_norm": 0.3514528274536133,
"learning_rate": 0.0001497996805658238,
"loss": 1.1257,
"step": 26
},
{
"epoch": 0.3375,
"grad_norm": 0.42414528131484985,
"learning_rate": 0.00014901875087822337,
"loss": 1.1463,
"step": 27
},
{
"epoch": 0.35,
"grad_norm": 0.35511669516563416,
"learning_rate": 0.0001482112131483274,
"loss": 1.141,
"step": 28
},
{
"epoch": 0.3625,
"grad_norm": 0.3799460530281067,
"learning_rate": 0.00014737737869730292,
"loss": 1.1414,
"step": 29
},
{
"epoch": 0.375,
"grad_norm": 0.26333189010620117,
"learning_rate": 0.00014651756898420365,
"loss": 1.1352,
"step": 30
},
{
"epoch": 0.3875,
"grad_norm": 0.37996864318847656,
"learning_rate": 0.0001456321154820411,
"loss": 1.1167,
"step": 31
},
{
"epoch": 0.4,
"grad_norm": 0.3210310637950897,
"learning_rate": 0.00014472135954999581,
"loss": 1.113,
"step": 32
},
{
"epoch": 0.4125,
"grad_norm": 0.342960923910141,
"learning_rate": 0.00014378565230181657,
"loss": 1.1201,
"step": 33
},
{
"epoch": 0.425,
"grad_norm": 0.30171331763267517,
"learning_rate": 0.0001428253544704596,
"loss": 1.1303,
"step": 34
},
{
"epoch": 0.4375,
"grad_norm": 0.3308579623699188,
"learning_rate": 0.00014184083626901897,
"loss": 1.135,
"step": 35
},
{
"epoch": 0.45,
"grad_norm": 0.33749139308929443,
"learning_rate": 0.0001408324772480025,
"loss": 1.1413,
"step": 36
},
{
"epoch": 0.4625,
"grad_norm": 0.29873886704444885,
"learning_rate": 0.00013980066614900776,
"loss": 1.1406,
"step": 37
},
{
"epoch": 0.475,
"grad_norm": 0.25276514887809753,
"learning_rate": 0.00013874580075485485,
"loss": 1.1421,
"step": 38
},
{
"epoch": 0.4875,
"grad_norm": 0.2849913537502289,
"learning_rate": 0.00013766828773623352,
"loss": 1.1298,
"step": 39
},
{
"epoch": 0.5,
"grad_norm": 0.27665936946868896,
"learning_rate": 0.00013656854249492382,
"loss": 1.1052,
"step": 40
},
{
"epoch": 0.5125,
"grad_norm": 0.31618547439575195,
"learning_rate": 0.0001354469890036509,
"loss": 1.1124,
"step": 41
},
{
"epoch": 0.525,
"grad_norm": 0.30855098366737366,
"learning_rate": 0.00013430405964263536,
"loss": 1.1164,
"step": 42
},
{
"epoch": 0.5375,
"grad_norm": 0.24974325299263,
"learning_rate": 0.00013314019503290255,
"loss": 1.1379,
"step": 43
},
{
"epoch": 0.55,
"grad_norm": 0.259245365858078,
"learning_rate": 0.00013195584386641469,
"loss": 1.0963,
"step": 44
},
{
"epoch": 0.5625,
"grad_norm": 0.3342917561531067,
"learning_rate": 0.00013075146273309164,
"loss": 1.1089,
"step": 45
},
{
"epoch": 0.575,
"grad_norm": 0.3317720293998718,
"learning_rate": 0.00012952751594478675,
"loss": 1.1226,
"step": 46
},
{
"epoch": 0.5875,
"grad_norm": 0.2566727101802826,
"learning_rate": 0.0001282844753562857,
"loss": 1.1035,
"step": 47
},
{
"epoch": 0.6,
"grad_norm": 0.25012263655662537,
"learning_rate": 0.00012702282018339786,
"loss": 1.0713,
"step": 48
},
{
"epoch": 0.6125,
"grad_norm": 0.2855740189552307,
"learning_rate": 0.00012574303681820898,
"loss": 1.1232,
"step": 49
},
{
"epoch": 0.625,
"grad_norm": 0.21377117931842804,
"learning_rate": 0.0001244456186415682,
"loss": 1.0726,
"step": 50
},
{
"epoch": 0.6375,
"grad_norm": 0.3012278079986572,
"learning_rate": 0.00012313106583288004,
"loss": 1.0855,
"step": 51
},
{
"epoch": 0.65,
"grad_norm": 0.2754627764225006,
"learning_rate": 0.00012179988517727591,
"loss": 1.113,
"step": 52
},
{
"epoch": 0.6625,
"grad_norm": 0.2773728668689728,
"learning_rate": 0.00012045258987023879,
"loss": 1.0931,
"step": 53
},
{
"epoch": 0.675,
"grad_norm": 0.3616091012954712,
"learning_rate": 0.00011908969931975641,
"loss": 1.1007,
"step": 54
},
{
"epoch": 0.6875,
"grad_norm": 0.28011709451675415,
"learning_rate": 0.00011771173894607985,
"loss": 1.1312,
"step": 55
},
{
"epoch": 0.7,
"grad_norm": 0.19245974719524384,
"learning_rate": 0.00011631923997916375,
"loss": 1.0784,
"step": 56
},
{
"epoch": 0.7125,
"grad_norm": 0.2807864844799042,
"learning_rate": 0.00011491273925386736,
"loss": 1.0766,
"step": 57
},
{
"epoch": 0.725,
"grad_norm": 0.25869062542915344,
"learning_rate": 0.00011349277900299426,
"loss": 1.0929,
"step": 58
},
{
"epoch": 0.7375,
"grad_norm": 0.20374780893325806,
"learning_rate": 0.00011205990664825127,
"loss": 1.0977,
"step": 59
},
{
"epoch": 0.75,
"grad_norm": 0.275302529335022,
"learning_rate": 0.00011061467458920719,
"loss": 1.1218,
"step": 60
},
{
"epoch": 0.7625,
"grad_norm": 0.26479312777519226,
"learning_rate": 0.00010915763999033201,
"loss": 1.0972,
"step": 61
},
{
"epoch": 0.775,
"grad_norm": 0.20327049493789673,
"learning_rate": 0.00010768936456619945,
"loss": 1.0723,
"step": 62
},
{
"epoch": 0.7875,
"grad_norm": 0.18908362090587616,
"learning_rate": 0.0001062104143649355,
"loss": 1.1059,
"step": 63
},
{
"epoch": 0.8,
"grad_norm": 0.2153932750225067,
"learning_rate": 0.0001047213595499958,
"loss": 1.1112,
"step": 64
},
{
"epoch": 0.8125,
"grad_norm": 0.23449215292930603,
"learning_rate": 0.000103222774180357,
"loss": 1.1125,
"step": 65
},
{
"epoch": 0.825,
"grad_norm": 0.19739866256713867,
"learning_rate": 0.00010171523598920594,
"loss": 1.0506,
"step": 66
},
{
"epoch": 0.8375,
"grad_norm": 0.579247236251831,
"learning_rate": 0.00010019932616121264,
"loss": 1.0599,
"step": 67
},
{
"epoch": 0.85,
"grad_norm": 0.2158878892660141,
"learning_rate": 9.867562910847246e-05,
"loss": 1.1116,
"step": 68
},
{
"epoch": 0.8625,
"grad_norm": 0.24534355103969574,
"learning_rate": 9.714473224520406e-05,
"loss": 1.1039,
"step": 69
},
{
"epoch": 0.875,
"grad_norm": 0.1604059487581253,
"learning_rate": 9.560722576129029e-05,
"loss": 1.1007,
"step": 70
},
{
"epoch": 0.8875,
"grad_norm": 0.24135896563529968,
"learning_rate": 9.406370239474839e-05,
"loss": 1.0976,
"step": 71
},
{
"epoch": 0.9,
"grad_norm": 0.2200448215007782,
"learning_rate": 9.251475720321848e-05,
"loss": 1.1001,
"step": 72
},
{
"epoch": 0.9125,
"grad_norm": 0.17517372965812683,
"learning_rate": 9.096098733455746e-05,
"loss": 1.0864,
"step": 73
},
{
"epoch": 0.925,
"grad_norm": 0.23631267249584198,
"learning_rate": 8.940299179662703e-05,
"loss": 1.0976,
"step": 74
},
{
"epoch": 0.9375,
"grad_norm": 0.17627741396427155,
"learning_rate": 8.784137122636488e-05,
"loss": 1.1049,
"step": 75
},
{
"epoch": 0.95,
"grad_norm": 0.1840021312236786,
"learning_rate": 8.627672765822762e-05,
"loss": 1.0504,
"step": 76
},
{
"epoch": 0.9625,
"grad_norm": 0.1925836205482483,
"learning_rate": 8.470966429209512e-05,
"loss": 1.1028,
"step": 77
},
{
"epoch": 0.975,
"grad_norm": 0.18122681975364685,
"learning_rate": 8.31407852607255e-05,
"loss": 1.0605,
"step": 78
},
{
"epoch": 0.9875,
"grad_norm": 0.16970321536064148,
"learning_rate": 8.157069539685026e-05,
"loss": 1.1069,
"step": 79
},
{
"epoch": 1.0,
"grad_norm": 0.2044173628091812,
"learning_rate": 8e-05,
"loss": 1.1173,
"step": 80
},
{
"epoch": 1.0125,
"grad_norm": 0.27160316705703735,
"learning_rate": 7.842930460314975e-05,
"loss": 1.0536,
"step": 81
},
{
"epoch": 1.025,
"grad_norm": 0.46974608302116394,
"learning_rate": 7.685921473927454e-05,
"loss": 1.1236,
"step": 82
},
{
"epoch": 1.0375,
"grad_norm": 0.4345127046108246,
"learning_rate": 7.529033570790488e-05,
"loss": 1.0927,
"step": 83
},
{
"epoch": 1.05,
"grad_norm": 0.3712212145328522,
"learning_rate": 7.372327234177242e-05,
"loss": 1.0861,
"step": 84
},
{
"epoch": 1.0625,
"grad_norm": 0.3404874801635742,
"learning_rate": 7.215862877363515e-05,
"loss": 1.108,
"step": 85
},
{
"epoch": 1.075,
"grad_norm": 0.411827951669693,
"learning_rate": 7.0597008203373e-05,
"loss": 1.0958,
"step": 86
},
{
"epoch": 1.0875,
"grad_norm": 0.28986656665802,
"learning_rate": 6.903901266544258e-05,
"loss": 1.0456,
"step": 87
},
{
"epoch": 1.1,
"grad_norm": 0.35029786825180054,
"learning_rate": 6.748524279678152e-05,
"loss": 1.1224,
"step": 88
},
{
"epoch": 1.1125,
"grad_norm": 0.2584652900695801,
"learning_rate": 6.593629760525164e-05,
"loss": 1.0932,
"step": 89
},
{
"epoch": 1.125,
"grad_norm": 0.30658528208732605,
"learning_rate": 6.439277423870975e-05,
"loss": 1.1091,
"step": 90
},
{
"epoch": 1.1375,
"grad_norm": 0.2145494520664215,
"learning_rate": 6.285526775479596e-05,
"loss": 1.0921,
"step": 91
},
{
"epoch": 1.15,
"grad_norm": 0.2773503065109253,
"learning_rate": 6.13243708915276e-05,
"loss": 1.0642,
"step": 92
},
{
"epoch": 1.1625,
"grad_norm": 0.2383718639612198,
"learning_rate": 5.9800673838787364e-05,
"loss": 1.0832,
"step": 93
},
{
"epoch": 1.175,
"grad_norm": 0.23632678389549255,
"learning_rate": 5.828476401079407e-05,
"loss": 1.0666,
"step": 94
},
{
"epoch": 1.1875,
"grad_norm": 0.21097873151302338,
"learning_rate": 5.677722581964303e-05,
"loss": 1.0342,
"step": 95
},
{
"epoch": 1.2,
"grad_norm": 0.20526330173015594,
"learning_rate": 5.5278640450004216e-05,
"loss": 1.0668,
"step": 96
},
{
"epoch": 1.2125,
"grad_norm": 0.19396792352199554,
"learning_rate": 5.3789585635064534e-05,
"loss": 1.0359,
"step": 97
},
{
"epoch": 1.225,
"grad_norm": 0.18029625713825226,
"learning_rate": 5.231063543380055e-05,
"loss": 1.0305,
"step": 98
},
{
"epoch": 1.2375,
"grad_norm": 0.16607148945331573,
"learning_rate": 5.084236000966803e-05,
"loss": 1.0516,
"step": 99
},
{
"epoch": 1.25,
"grad_norm": 0.1904178261756897,
"learning_rate": 4.9385325410792824e-05,
"loss": 1.0456,
"step": 100
},
{
"epoch": 1.2625,
"grad_norm": 0.16171352565288544,
"learning_rate": 4.794009335174874e-05,
"loss": 1.0482,
"step": 101
},
{
"epoch": 1.275,
"grad_norm": 0.17175588011741638,
"learning_rate": 4.650722099700578e-05,
"loss": 1.0442,
"step": 102
},
{
"epoch": 1.2875,
"grad_norm": 0.147624671459198,
"learning_rate": 4.508726074613262e-05,
"loss": 1.0694,
"step": 103
},
{
"epoch": 1.3,
"grad_norm": 0.168911874294281,
"learning_rate": 4.3680760020836266e-05,
"loss": 1.0556,
"step": 104
},
{
"epoch": 1.3125,
"grad_norm": 0.15126970410346985,
"learning_rate": 4.2288261053920186e-05,
"loss": 1.044,
"step": 105
},
{
"epoch": 1.325,
"grad_norm": 0.15997202694416046,
"learning_rate": 4.0910300680243636e-05,
"loss": 1.0231,
"step": 106
},
{
"epoch": 1.3375,
"grad_norm": 0.16348402202129364,
"learning_rate": 3.954741012976125e-05,
"loss": 1.0468,
"step": 107
},
{
"epoch": 1.35,
"grad_norm": 0.16921056807041168,
"learning_rate": 3.8200114822724096e-05,
"loss": 1.0452,
"step": 108
},
{
"epoch": 1.3625,
"grad_norm": 0.1568213850259781,
"learning_rate": 3.686893416711998e-05,
"loss": 1.0487,
"step": 109
},
{
"epoch": 1.375,
"grad_norm": 0.12868821620941162,
"learning_rate": 3.5554381358431845e-05,
"loss": 1.0447,
"step": 110
},
{
"epoch": 1.3875,
"grad_norm": 0.14700743556022644,
"learning_rate": 3.425696318179103e-05,
"loss": 1.0288,
"step": 111
},
{
"epoch": 1.4,
"grad_norm": 0.1439713090658188,
"learning_rate": 3.297717981660216e-05,
"loss": 1.0269,
"step": 112
},
{
"epoch": 1.4125,
"grad_norm": 0.12184255570173264,
"learning_rate": 3.1715524643714286e-05,
"loss": 1.0361,
"step": 113
},
{
"epoch": 1.425,
"grad_norm": 0.14732472598552704,
"learning_rate": 3.0472484055213276e-05,
"loss": 1.0476,
"step": 114
},
{
"epoch": 1.4375,
"grad_norm": 0.14000268280506134,
"learning_rate": 2.9248537266908373e-05,
"loss": 1.0531,
"step": 115
},
{
"epoch": 1.45,
"grad_norm": 0.12124885618686676,
"learning_rate": 2.804415613358532e-05,
"loss": 1.0611,
"step": 116
},
{
"epoch": 1.4625,
"grad_norm": 0.13823653757572174,
"learning_rate": 2.685980496709749e-05,
"loss": 1.0625,
"step": 117
},
{
"epoch": 1.475,
"grad_norm": 0.1250331848859787,
"learning_rate": 2.569594035736466e-05,
"loss": 1.0647,
"step": 118
},
{
"epoch": 1.4875,
"grad_norm": 0.11904938519001007,
"learning_rate": 2.4553010996349143e-05,
"loss": 1.0543,
"step": 119
},
{
"epoch": 1.5,
"grad_norm": 0.12068229168653488,
"learning_rate": 2.3431457505076205e-05,
"loss": 1.0325,
"step": 120
},
{
"epoch": 1.5125,
"grad_norm": 0.12019164115190506,
"learning_rate": 2.2331712263766495e-05,
"loss": 1.041,
"step": 121
},
{
"epoch": 1.525,
"grad_norm": 0.12757574021816254,
"learning_rate": 2.1254199245145177e-05,
"loss": 1.0433,
"step": 122
},
{
"epoch": 1.5375,
"grad_norm": 0.12285693734884262,
"learning_rate": 2.0199333850992245e-05,
"loss": 1.0657,
"step": 123
},
{
"epoch": 1.55,
"grad_norm": 0.12083045393228531,
"learning_rate": 1.9167522751997527e-05,
"loss": 1.0292,
"step": 124
},
{
"epoch": 1.5625,
"grad_norm": 0.11348845809698105,
"learning_rate": 1.815916373098104e-05,
"loss": 1.04,
"step": 125
},
{
"epoch": 1.575,
"grad_norm": 0.12212307006120682,
"learning_rate": 1.7174645529540424e-05,
"loss": 1.0556,
"step": 126
},
{
"epoch": 1.5875,
"grad_norm": 0.1128983274102211,
"learning_rate": 1.621434769818344e-05,
"loss": 1.0365,
"step": 127
},
{
"epoch": 1.6,
"grad_norm": 0.11498909443616867,
"learning_rate": 1.5278640450004213e-05,
"loss": 1.0072,
"step": 128
},
{
"epoch": 1.6125,
"grad_norm": 0.12018037587404251,
"learning_rate": 1.4367884517958914e-05,
"loss": 1.0589,
"step": 129
},
{
"epoch": 1.625,
"grad_norm": 0.10916531831026077,
"learning_rate": 1.3482431015796373e-05,
"loss": 1.0112,
"step": 130
},
{
"epoch": 1.6375,
"grad_norm": 0.10843243449926376,
"learning_rate": 1.2622621302697087e-05,
"loss": 1.0249,
"step": 131
},
{
"epoch": 1.65,
"grad_norm": 0.111574187874794,
"learning_rate": 1.1788786851672628e-05,
"loss": 1.0507,
"step": 132
},
{
"epoch": 1.6625,
"grad_norm": 0.10717642307281494,
"learning_rate": 1.0981249121776654e-05,
"loss": 1.0328,
"step": 133
},
{
"epoch": 1.675,
"grad_norm": 0.11177720874547958,
"learning_rate": 1.0200319434176227e-05,
"loss": 1.0412,
"step": 134
},
{
"epoch": 1.6875,
"grad_norm": 0.10437416285276413,
"learning_rate": 9.446298852131605e-06,
"loss": 1.072,
"step": 135
},
{
"epoch": 1.7,
"grad_norm": 0.10602446645498276,
"learning_rate": 8.719478064930578e-06,
"loss": 1.022,
"step": 136
},
{
"epoch": 1.7125,
"grad_norm": 0.10757151246070862,
"learning_rate": 8.020137275822297e-06,
"loss": 1.021,
"step": 137
},
{
"epoch": 1.725,
"grad_norm": 0.10059204697608948,
"learning_rate": 7.348546093993492e-06,
"loss": 1.0369,
"step": 138
},
{
"epoch": 1.7375,
"grad_norm": 0.09817243367433548,
"learning_rate": 6.704963430629132e-06,
"loss": 1.0433,
"step": 139
},
{
"epoch": 1.75,
"grad_norm": 1.0110856294631958,
"learning_rate": 6.0896373990970614e-06,
"loss": 1.0893,
"step": 140
},
{
"epoch": 1.7625,
"grad_norm": 0.1014508381485939,
"learning_rate": 5.502805219295715e-06,
"loss": 1.0433,
"step": 141
},
{
"epoch": 1.775,
"grad_norm": 0.09631301462650299,
"learning_rate": 4.944693126201276e-06,
"loss": 1.0207,
"step": 142
},
{
"epoch": 1.7875,
"grad_norm": 0.09827422350645065,
"learning_rate": 4.415516282649756e-06,
"loss": 1.0538,
"step": 143
},
{
"epoch": 1.8,
"grad_norm": 0.100394107401371,
"learning_rate": 3.915478696387718e-06,
"loss": 1.0606,
"step": 144
},
{
"epoch": 1.8125,
"grad_norm": 0.0930555984377861,
"learning_rate": 3.4447731414232945e-06,
"loss": 1.0614,
"step": 145
},
{
"epoch": 1.825,
"grad_norm": 0.09103138744831085,
"learning_rate": 3.0035810837082267e-06,
"loss": 1.0028,
"step": 146
},
{
"epoch": 1.8375,
"grad_norm": 0.09185432642698288,
"learning_rate": 2.5920726111792195e-06,
"loss": 1.0137,
"step": 147
},
{
"epoch": 1.85,
"grad_norm": 0.09481006115674973,
"learning_rate": 2.2104063681858757e-06,
"loss": 1.0631,
"step": 148
},
{
"epoch": 1.8625,
"grad_norm": 0.08825208991765976,
"learning_rate": 1.8587294943302092e-06,
"loss": 1.0562,
"step": 149
},
{
"epoch": 1.875,
"grad_norm": 0.09086597710847855,
"learning_rate": 1.5371775677415656e-06,
"loss": 1.0534,
"step": 150
},
{
"epoch": 1.8875,
"grad_norm": 0.09239210933446884,
"learning_rate": 1.2458745528086723e-06,
"loss": 1.0517,
"step": 151
},
{
"epoch": 1.9,
"grad_norm": 0.091704361140728,
"learning_rate": 9.849327523889873e-07,
"loss": 1.0547,
"step": 152
},
{
"epoch": 1.9125,
"grad_norm": 0.0930432379245758,
"learning_rate": 7.5445276451382e-07,
"loss": 1.0418,
"step": 153
},
{
"epoch": 1.925,
"grad_norm": 0.08982589095830917,
"learning_rate": 5.545234436058966e-07,
"loss": 1.0541,
"step": 154
},
{
"epoch": 1.9375,
"grad_norm": 0.09276595711708069,
"learning_rate": 3.852218662242546e-07,
"loss": 1.0613,
"step": 155
},
{
"epoch": 1.95,
"grad_norm": 0.08679981529712677,
"learning_rate": 2.466133013497629e-07,
"loss": 1.0097,
"step": 156
},
{
"epoch": 1.9625,
"grad_norm": 0.08747802674770355,
"learning_rate": 1.3875118522273412e-07,
"loss": 1.0606,
"step": 157
},
{
"epoch": 1.975,
"grad_norm": 0.08833891898393631,
"learning_rate": 6.167710074216792e-08,
"loss": 1.0207,
"step": 158
},
{
"epoch": 1.9875,
"grad_norm": 0.0912172719836235,
"learning_rate": 1.5420761434814523e-08,
"loss": 1.066,
"step": 159
},
{
"epoch": 2.0,
"grad_norm": 0.09457752853631973,
"learning_rate": 0.0,
"loss": 1.0772,
"step": 160
}
],
"logging_steps": 1,
"max_steps": 160,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 40,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.276128179356959e+18,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}