{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 144, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.020942408376963352, "grad_norm": 0.8203125, "learning_rate": 0.0, "loss": 2.3469, "step": 1 }, { "epoch": 0.041884816753926704, "grad_norm": 0.86328125, "learning_rate": 4.000000000000001e-06, "loss": 2.3465, "step": 2 }, { "epoch": 0.06282722513089005, "grad_norm": 0.828125, "learning_rate": 8.000000000000001e-06, "loss": 2.3569, "step": 3 }, { "epoch": 0.08376963350785341, "grad_norm": 0.89453125, "learning_rate": 1.2e-05, "loss": 2.3221, "step": 4 }, { "epoch": 0.10471204188481675, "grad_norm": 0.765625, "learning_rate": 1.6000000000000003e-05, "loss": 2.3378, "step": 5 }, { "epoch": 0.1256544502617801, "grad_norm": 0.8125, "learning_rate": 2e-05, "loss": 2.372, "step": 6 }, { "epoch": 0.14659685863874344, "grad_norm": 0.7578125, "learning_rate": 1.985611510791367e-05, "loss": 2.3172, "step": 7 }, { "epoch": 0.16753926701570682, "grad_norm": 0.703125, "learning_rate": 1.971223021582734e-05, "loss": 2.3166, "step": 8 }, { "epoch": 0.18848167539267016, "grad_norm": 0.7578125, "learning_rate": 1.956834532374101e-05, "loss": 2.3642, "step": 9 }, { "epoch": 0.2094240837696335, "grad_norm": 0.69921875, "learning_rate": 1.9424460431654675e-05, "loss": 2.4136, "step": 10 }, { "epoch": 0.23036649214659685, "grad_norm": 0.66796875, "learning_rate": 1.9280575539568347e-05, "loss": 2.2907, "step": 11 }, { "epoch": 0.2513089005235602, "grad_norm": 0.63671875, "learning_rate": 1.9136690647482016e-05, "loss": 2.1985, "step": 12 }, { "epoch": 0.27225130890052357, "grad_norm": 0.61328125, "learning_rate": 1.8992805755395684e-05, "loss": 2.2085, "step": 13 }, { "epoch": 0.2931937172774869, "grad_norm": 0.59765625, "learning_rate": 1.8848920863309356e-05, "loss": 2.3054, "step": 14 }, { "epoch": 0.31413612565445026, "grad_norm": 0.56640625, "learning_rate": 1.870503597122302e-05, "loss": 2.1966, "step": 15 }, { "epoch": 0.33507853403141363, "grad_norm": 0.58203125, "learning_rate": 1.8561151079136693e-05, "loss": 2.1649, "step": 16 }, { "epoch": 0.35602094240837695, "grad_norm": 0.55859375, "learning_rate": 1.841726618705036e-05, "loss": 2.2583, "step": 17 }, { "epoch": 0.3769633507853403, "grad_norm": 0.5859375, "learning_rate": 1.827338129496403e-05, "loss": 2.3205, "step": 18 }, { "epoch": 0.39790575916230364, "grad_norm": 0.54296875, "learning_rate": 1.8129496402877698e-05, "loss": 2.1887, "step": 19 }, { "epoch": 0.418848167539267, "grad_norm": 0.5703125, "learning_rate": 1.7985611510791367e-05, "loss": 2.2294, "step": 20 }, { "epoch": 0.4397905759162304, "grad_norm": 0.578125, "learning_rate": 1.784172661870504e-05, "loss": 2.1661, "step": 21 }, { "epoch": 0.4607329842931937, "grad_norm": 0.5546875, "learning_rate": 1.7697841726618707e-05, "loss": 2.0935, "step": 22 }, { "epoch": 0.4816753926701571, "grad_norm": 0.52734375, "learning_rate": 1.7553956834532375e-05, "loss": 2.1805, "step": 23 }, { "epoch": 0.5026178010471204, "grad_norm": 0.6796875, "learning_rate": 1.7410071942446044e-05, "loss": 2.2131, "step": 24 }, { "epoch": 0.5235602094240838, "grad_norm": 0.53125, "learning_rate": 1.7266187050359712e-05, "loss": 2.175, "step": 25 }, { "epoch": 0.5445026178010471, "grad_norm": 0.50390625, "learning_rate": 1.7122302158273384e-05, "loss": 2.1707, "step": 26 }, { "epoch": 0.5654450261780105, "grad_norm": 0.51171875, "learning_rate": 1.6978417266187053e-05, "loss": 2.0651, "step": 27 }, { "epoch": 0.5863874345549738, "grad_norm": 0.53515625, "learning_rate": 1.683453237410072e-05, "loss": 2.1611, "step": 28 }, { "epoch": 0.6073298429319371, "grad_norm": 0.53125, "learning_rate": 1.669064748201439e-05, "loss": 2.1401, "step": 29 }, { "epoch": 0.6282722513089005, "grad_norm": 0.5390625, "learning_rate": 1.6546762589928058e-05, "loss": 2.1536, "step": 30 }, { "epoch": 0.6492146596858639, "grad_norm": 0.5234375, "learning_rate": 1.640287769784173e-05, "loss": 2.207, "step": 31 }, { "epoch": 0.6701570680628273, "grad_norm": 0.5234375, "learning_rate": 1.6258992805755398e-05, "loss": 2.1138, "step": 32 }, { "epoch": 0.6910994764397905, "grad_norm": 0.5078125, "learning_rate": 1.6115107913669067e-05, "loss": 2.1441, "step": 33 }, { "epoch": 0.7120418848167539, "grad_norm": 0.5, "learning_rate": 1.5971223021582735e-05, "loss": 2.1523, "step": 34 }, { "epoch": 0.7329842931937173, "grad_norm": 0.50390625, "learning_rate": 1.5827338129496403e-05, "loss": 2.1216, "step": 35 }, { "epoch": 0.7539267015706806, "grad_norm": 0.486328125, "learning_rate": 1.5683453237410072e-05, "loss": 2.0619, "step": 36 }, { "epoch": 0.774869109947644, "grad_norm": 0.498046875, "learning_rate": 1.5539568345323744e-05, "loss": 2.0156, "step": 37 }, { "epoch": 0.7958115183246073, "grad_norm": 0.50390625, "learning_rate": 1.5395683453237412e-05, "loss": 2.1156, "step": 38 }, { "epoch": 0.8167539267015707, "grad_norm": 0.5, "learning_rate": 1.525179856115108e-05, "loss": 2.0561, "step": 39 }, { "epoch": 0.837696335078534, "grad_norm": 0.52734375, "learning_rate": 1.5107913669064749e-05, "loss": 2.1226, "step": 40 }, { "epoch": 0.8586387434554974, "grad_norm": 0.51171875, "learning_rate": 1.496402877697842e-05, "loss": 2.1199, "step": 41 }, { "epoch": 0.8795811518324608, "grad_norm": 0.52734375, "learning_rate": 1.4820143884892086e-05, "loss": 2.1843, "step": 42 }, { "epoch": 0.900523560209424, "grad_norm": 0.515625, "learning_rate": 1.4676258992805756e-05, "loss": 2.103, "step": 43 }, { "epoch": 0.9214659685863874, "grad_norm": 0.53515625, "learning_rate": 1.4532374100719426e-05, "loss": 2.1346, "step": 44 }, { "epoch": 0.9424083769633508, "grad_norm": 0.5390625, "learning_rate": 1.4388489208633095e-05, "loss": 2.0813, "step": 45 }, { "epoch": 0.9633507853403142, "grad_norm": 0.55859375, "learning_rate": 1.4244604316546765e-05, "loss": 2.1774, "step": 46 }, { "epoch": 0.9842931937172775, "grad_norm": 0.51953125, "learning_rate": 1.4100719424460432e-05, "loss": 2.0659, "step": 47 }, { "epoch": 1.0, "grad_norm": 0.59765625, "learning_rate": 1.3956834532374102e-05, "loss": 2.0783, "step": 48 }, { "epoch": 1.0, "eval_loss": 2.0860884189605713, "eval_model_preparation_time": 0.0149, "eval_runtime": 8.1703, "eval_samples_per_second": 10.404, "eval_steps_per_second": 5.263, "step": 48 }, { "epoch": 1.0209424083769634, "grad_norm": 0.490234375, "learning_rate": 1.3812949640287772e-05, "loss": 2.0715, "step": 49 }, { "epoch": 1.0418848167539267, "grad_norm": 0.49609375, "learning_rate": 1.3669064748201439e-05, "loss": 2.0527, "step": 50 }, { "epoch": 1.0628272251308901, "grad_norm": 0.51171875, "learning_rate": 1.3525179856115109e-05, "loss": 1.9961, "step": 51 }, { "epoch": 1.0837696335078535, "grad_norm": 0.54296875, "learning_rate": 1.3381294964028777e-05, "loss": 2.1151, "step": 52 }, { "epoch": 1.1047120418848166, "grad_norm": 0.51171875, "learning_rate": 1.3237410071942447e-05, "loss": 2.0206, "step": 53 }, { "epoch": 1.12565445026178, "grad_norm": 0.515625, "learning_rate": 1.3093525179856117e-05, "loss": 2.1006, "step": 54 }, { "epoch": 1.1465968586387434, "grad_norm": 0.50390625, "learning_rate": 1.2949640287769784e-05, "loss": 2.0258, "step": 55 }, { "epoch": 1.1675392670157068, "grad_norm": 0.53125, "learning_rate": 1.2805755395683454e-05, "loss": 1.981, "step": 56 }, { "epoch": 1.1884816753926701, "grad_norm": 0.54296875, "learning_rate": 1.2661870503597123e-05, "loss": 2.1331, "step": 57 }, { "epoch": 1.2094240837696335, "grad_norm": 0.494140625, "learning_rate": 1.2517985611510793e-05, "loss": 2.1064, "step": 58 }, { "epoch": 1.2303664921465969, "grad_norm": 0.54296875, "learning_rate": 1.2374100719424463e-05, "loss": 1.9746, "step": 59 }, { "epoch": 1.2513089005235603, "grad_norm": 0.4921875, "learning_rate": 1.223021582733813e-05, "loss": 2.0598, "step": 60 }, { "epoch": 1.2722513089005236, "grad_norm": 0.53125, "learning_rate": 1.20863309352518e-05, "loss": 2.0676, "step": 61 }, { "epoch": 1.2931937172774868, "grad_norm": 0.5078125, "learning_rate": 1.1942446043165468e-05, "loss": 2.0734, "step": 62 }, { "epoch": 1.3141361256544504, "grad_norm": 0.5234375, "learning_rate": 1.1798561151079137e-05, "loss": 1.968, "step": 63 }, { "epoch": 1.3350785340314135, "grad_norm": 0.546875, "learning_rate": 1.1654676258992807e-05, "loss": 2.0268, "step": 64 }, { "epoch": 1.356020942408377, "grad_norm": 0.5390625, "learning_rate": 1.1510791366906475e-05, "loss": 2.0467, "step": 65 }, { "epoch": 1.3769633507853403, "grad_norm": 0.51171875, "learning_rate": 1.1366906474820146e-05, "loss": 2.0148, "step": 66 }, { "epoch": 1.3979057591623036, "grad_norm": 0.54296875, "learning_rate": 1.1223021582733812e-05, "loss": 2.0084, "step": 67 }, { "epoch": 1.418848167539267, "grad_norm": 0.51171875, "learning_rate": 1.1079136690647482e-05, "loss": 2.005, "step": 68 }, { "epoch": 1.4397905759162304, "grad_norm": 0.490234375, "learning_rate": 1.0935251798561153e-05, "loss": 2.0613, "step": 69 }, { "epoch": 1.4607329842931938, "grad_norm": 0.51171875, "learning_rate": 1.0791366906474821e-05, "loss": 2.0394, "step": 70 }, { "epoch": 1.4816753926701571, "grad_norm": 0.50390625, "learning_rate": 1.0647482014388491e-05, "loss": 2.0328, "step": 71 }, { "epoch": 1.5026178010471205, "grad_norm": 0.51953125, "learning_rate": 1.0503597122302158e-05, "loss": 2.0529, "step": 72 }, { "epoch": 1.5235602094240837, "grad_norm": 0.51953125, "learning_rate": 1.0359712230215828e-05, "loss": 1.9776, "step": 73 }, { "epoch": 1.5445026178010473, "grad_norm": 0.53515625, "learning_rate": 1.0215827338129498e-05, "loss": 2.0813, "step": 74 }, { "epoch": 1.5654450261780104, "grad_norm": 0.51953125, "learning_rate": 1.0071942446043167e-05, "loss": 1.9543, "step": 75 }, { "epoch": 1.5863874345549738, "grad_norm": 0.51171875, "learning_rate": 9.928057553956835e-06, "loss": 2.0868, "step": 76 }, { "epoch": 1.6073298429319371, "grad_norm": 0.51171875, "learning_rate": 9.784172661870505e-06, "loss": 2.0053, "step": 77 }, { "epoch": 1.6282722513089005, "grad_norm": 0.609375, "learning_rate": 9.640287769784174e-06, "loss": 2.074, "step": 78 }, { "epoch": 1.649214659685864, "grad_norm": 0.52734375, "learning_rate": 9.496402877697842e-06, "loss": 1.9674, "step": 79 }, { "epoch": 1.6701570680628273, "grad_norm": 0.5234375, "learning_rate": 9.35251798561151e-06, "loss": 2.0277, "step": 80 }, { "epoch": 1.6910994764397906, "grad_norm": 0.50390625, "learning_rate": 9.20863309352518e-06, "loss": 1.9616, "step": 81 }, { "epoch": 1.7120418848167538, "grad_norm": 0.5234375, "learning_rate": 9.064748201438849e-06, "loss": 2.0238, "step": 82 }, { "epoch": 1.7329842931937174, "grad_norm": 0.5234375, "learning_rate": 8.92086330935252e-06, "loss": 1.9599, "step": 83 }, { "epoch": 1.7539267015706805, "grad_norm": 0.54296875, "learning_rate": 8.776978417266188e-06, "loss": 2.033, "step": 84 }, { "epoch": 1.7748691099476441, "grad_norm": 0.5078125, "learning_rate": 8.633093525179856e-06, "loss": 2.0615, "step": 85 }, { "epoch": 1.7958115183246073, "grad_norm": 0.53125, "learning_rate": 8.489208633093526e-06, "loss": 1.9893, "step": 86 }, { "epoch": 1.8167539267015707, "grad_norm": 0.5546875, "learning_rate": 8.345323741007195e-06, "loss": 2.0189, "step": 87 }, { "epoch": 1.837696335078534, "grad_norm": 0.50390625, "learning_rate": 8.201438848920865e-06, "loss": 1.9337, "step": 88 }, { "epoch": 1.8586387434554974, "grad_norm": 0.53515625, "learning_rate": 8.057553956834533e-06, "loss": 2.0259, "step": 89 }, { "epoch": 1.8795811518324608, "grad_norm": 0.51171875, "learning_rate": 7.913669064748202e-06, "loss": 2.0442, "step": 90 }, { "epoch": 1.900523560209424, "grad_norm": 0.52734375, "learning_rate": 7.769784172661872e-06, "loss": 2.0282, "step": 91 }, { "epoch": 1.9214659685863875, "grad_norm": 0.5234375, "learning_rate": 7.62589928057554e-06, "loss": 2.0452, "step": 92 }, { "epoch": 1.9424083769633507, "grad_norm": 0.5234375, "learning_rate": 7.48201438848921e-06, "loss": 2.0285, "step": 93 }, { "epoch": 1.9633507853403143, "grad_norm": 0.5234375, "learning_rate": 7.338129496402878e-06, "loss": 1.9695, "step": 94 }, { "epoch": 1.9842931937172774, "grad_norm": 0.5234375, "learning_rate": 7.194244604316547e-06, "loss": 2.0677, "step": 95 }, { "epoch": 2.0, "grad_norm": 0.62890625, "learning_rate": 7.050359712230216e-06, "loss": 2.0551, "step": 96 }, { "epoch": 2.0, "eval_loss": 2.038198471069336, "eval_model_preparation_time": 0.0149, "eval_runtime": 8.0804, "eval_samples_per_second": 10.519, "eval_steps_per_second": 5.322, "step": 96 }, { "epoch": 2.020942408376963, "grad_norm": 0.54296875, "learning_rate": 6.906474820143886e-06, "loss": 2.0502, "step": 97 }, { "epoch": 2.0418848167539267, "grad_norm": 0.5078125, "learning_rate": 6.762589928057554e-06, "loss": 2.0383, "step": 98 }, { "epoch": 2.06282722513089, "grad_norm": 0.515625, "learning_rate": 6.618705035971224e-06, "loss": 2.04, "step": 99 }, { "epoch": 2.0837696335078535, "grad_norm": 0.54296875, "learning_rate": 6.474820143884892e-06, "loss": 1.9828, "step": 100 }, { "epoch": 2.1047120418848166, "grad_norm": 0.50390625, "learning_rate": 6.330935251798561e-06, "loss": 2.025, "step": 101 }, { "epoch": 2.1256544502617802, "grad_norm": 0.5234375, "learning_rate": 6.1870503597122315e-06, "loss": 1.9143, "step": 102 }, { "epoch": 2.1465968586387434, "grad_norm": 0.50390625, "learning_rate": 6.0431654676259e-06, "loss": 1.998, "step": 103 }, { "epoch": 2.167539267015707, "grad_norm": 0.5234375, "learning_rate": 5.899280575539568e-06, "loss": 2.0212, "step": 104 }, { "epoch": 2.18848167539267, "grad_norm": 0.53125, "learning_rate": 5.755395683453238e-06, "loss": 2.0234, "step": 105 }, { "epoch": 2.2094240837696333, "grad_norm": 0.5234375, "learning_rate": 5.611510791366906e-06, "loss": 1.9175, "step": 106 }, { "epoch": 2.230366492146597, "grad_norm": 0.5234375, "learning_rate": 5.467625899280576e-06, "loss": 2.0098, "step": 107 }, { "epoch": 2.25130890052356, "grad_norm": 0.51953125, "learning_rate": 5.3237410071942456e-06, "loss": 1.9502, "step": 108 }, { "epoch": 2.2722513089005236, "grad_norm": 0.51953125, "learning_rate": 5.179856115107914e-06, "loss": 1.9713, "step": 109 }, { "epoch": 2.2931937172774868, "grad_norm": 0.546875, "learning_rate": 5.035971223021583e-06, "loss": 1.8965, "step": 110 }, { "epoch": 2.3141361256544504, "grad_norm": 0.5625, "learning_rate": 4.892086330935253e-06, "loss": 2.0483, "step": 111 }, { "epoch": 2.3350785340314135, "grad_norm": 0.5234375, "learning_rate": 4.748201438848921e-06, "loss": 1.9471, "step": 112 }, { "epoch": 2.356020942408377, "grad_norm": 0.53125, "learning_rate": 4.60431654676259e-06, "loss": 1.9803, "step": 113 }, { "epoch": 2.3769633507853403, "grad_norm": 0.53515625, "learning_rate": 4.46043165467626e-06, "loss": 2.0178, "step": 114 }, { "epoch": 2.3979057591623034, "grad_norm": 0.5390625, "learning_rate": 4.316546762589928e-06, "loss": 2.0112, "step": 115 }, { "epoch": 2.418848167539267, "grad_norm": 0.5703125, "learning_rate": 4.172661870503597e-06, "loss": 2.0413, "step": 116 }, { "epoch": 2.4397905759162306, "grad_norm": 0.52734375, "learning_rate": 4.028776978417267e-06, "loss": 2.011, "step": 117 }, { "epoch": 2.4607329842931938, "grad_norm": 0.51171875, "learning_rate": 3.884892086330936e-06, "loss": 2.0063, "step": 118 }, { "epoch": 2.481675392670157, "grad_norm": 0.5234375, "learning_rate": 3.741007194244605e-06, "loss": 1.9494, "step": 119 }, { "epoch": 2.5026178010471205, "grad_norm": 0.5859375, "learning_rate": 3.5971223021582737e-06, "loss": 2.1046, "step": 120 }, { "epoch": 2.5235602094240837, "grad_norm": 0.50390625, "learning_rate": 3.453237410071943e-06, "loss": 1.9551, "step": 121 }, { "epoch": 2.5445026178010473, "grad_norm": 0.53125, "learning_rate": 3.309352517985612e-06, "loss": 1.9959, "step": 122 }, { "epoch": 2.5654450261780104, "grad_norm": 0.53125, "learning_rate": 3.1654676258992807e-06, "loss": 1.9538, "step": 123 }, { "epoch": 2.5863874345549736, "grad_norm": 0.55078125, "learning_rate": 3.02158273381295e-06, "loss": 2.0185, "step": 124 }, { "epoch": 2.607329842931937, "grad_norm": 0.5390625, "learning_rate": 2.877697841726619e-06, "loss": 2.0775, "step": 125 }, { "epoch": 2.6282722513089007, "grad_norm": 0.54296875, "learning_rate": 2.733812949640288e-06, "loss": 1.9863, "step": 126 }, { "epoch": 2.649214659685864, "grad_norm": 0.5390625, "learning_rate": 2.589928057553957e-06, "loss": 1.969, "step": 127 }, { "epoch": 2.670157068062827, "grad_norm": 0.53515625, "learning_rate": 2.4460431654676263e-06, "loss": 2.0218, "step": 128 }, { "epoch": 2.6910994764397906, "grad_norm": 0.55078125, "learning_rate": 2.302158273381295e-06, "loss": 1.9308, "step": 129 }, { "epoch": 2.712041884816754, "grad_norm": 0.50390625, "learning_rate": 2.158273381294964e-06, "loss": 1.9536, "step": 130 }, { "epoch": 2.7329842931937174, "grad_norm": 0.5390625, "learning_rate": 2.0143884892086333e-06, "loss": 1.9376, "step": 131 }, { "epoch": 2.7539267015706805, "grad_norm": 0.71484375, "learning_rate": 1.8705035971223024e-06, "loss": 1.9802, "step": 132 }, { "epoch": 2.774869109947644, "grad_norm": 0.51953125, "learning_rate": 1.7266187050359715e-06, "loss": 1.9955, "step": 133 }, { "epoch": 2.7958115183246073, "grad_norm": 0.5234375, "learning_rate": 1.5827338129496403e-06, "loss": 1.9325, "step": 134 }, { "epoch": 2.816753926701571, "grad_norm": 0.5234375, "learning_rate": 1.4388489208633094e-06, "loss": 1.9576, "step": 135 }, { "epoch": 2.837696335078534, "grad_norm": 0.51171875, "learning_rate": 1.2949640287769785e-06, "loss": 1.9662, "step": 136 }, { "epoch": 2.858638743455497, "grad_norm": 0.52734375, "learning_rate": 1.1510791366906476e-06, "loss": 2.0159, "step": 137 }, { "epoch": 2.8795811518324608, "grad_norm": 0.5390625, "learning_rate": 1.0071942446043167e-06, "loss": 1.9664, "step": 138 }, { "epoch": 2.900523560209424, "grad_norm": 0.53125, "learning_rate": 8.633093525179857e-07, "loss": 1.9848, "step": 139 }, { "epoch": 2.9214659685863875, "grad_norm": 0.5, "learning_rate": 7.194244604316547e-07, "loss": 1.9385, "step": 140 }, { "epoch": 2.9424083769633507, "grad_norm": 0.5625, "learning_rate": 5.755395683453238e-07, "loss": 2.0412, "step": 141 }, { "epoch": 2.9633507853403143, "grad_norm": 0.53125, "learning_rate": 4.3165467625899287e-07, "loss": 2.0406, "step": 142 }, { "epoch": 2.9842931937172774, "grad_norm": 0.5234375, "learning_rate": 2.877697841726619e-07, "loss": 2.005, "step": 143 }, { "epoch": 3.0, "grad_norm": 0.6171875, "learning_rate": 1.4388489208633095e-07, "loss": 1.997, "step": 144 } ], "logging_steps": 1, "max_steps": 144, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 5000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.364548133830656e+16, "train_batch_size": 2, "trial_name": null, "trial_params": null }