Files
llama-3-8b-base-r-dpo-ultra…/trainer_state.json
ModelHub XC 8a58a992dc 初始化项目,由ModelHub XC社区提供模型
Model: jackf857/llama-3-8b-base-r-dpo-ultrafeedback-4xh200
Source: Original Platform
2026-04-25 12:06:49 +08:00

896 lines
31 KiB
JSON

{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9989528795811519,
"eval_steps": 200,
"global_step": 477,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0020942408376963353,
"grad_norm": 286.114501953125,
"learning_rate": 0.0,
"logits/chosen": -0.5995081663131714,
"logits/rejected": -0.6144353747367859,
"logps/chosen": -267.5272216796875,
"logps/ref_chosen": -267.5935363769531,
"logps/ref_rejected": -204.2306671142578,
"logps/rejected": -204.23907470703125,
"loss": 5.5602,
"r_dpo/chosen_len": 257.75,
"r_dpo/length_delta": 47.875,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 209.875,
"step": 1
},
{
"epoch": 0.020942408376963352,
"grad_norm": 265.7362365722656,
"learning_rate": 9.375e-08,
"logits/chosen": -0.6309213042259216,
"logits/rejected": -0.63545161485672,
"logps/chosen": -296.74920654296875,
"logps/ref_chosen": -296.72576904296875,
"logps/ref_rejected": -258.9681396484375,
"logps/rejected": -259.0067443847656,
"loss": 5.5477,
"r_dpo/chosen_len": 291.8680419921875,
"r_dpo/length_delta": 49.76388931274414,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 242.1041717529297,
"step": 10
},
{
"epoch": 0.041884816753926704,
"grad_norm": 296.5692138671875,
"learning_rate": 1.9791666666666664e-07,
"logits/chosen": -0.597561240196228,
"logits/rejected": -0.6285912990570068,
"logps/chosen": -297.9436340332031,
"logps/ref_chosen": -297.9349365234375,
"logps/ref_rejected": -256.9902648925781,
"logps/rejected": -256.98876953125,
"loss": 5.5369,
"r_dpo/chosen_len": 291.29998779296875,
"r_dpo/length_delta": 52.89374923706055,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 238.40625,
"step": 20
},
{
"epoch": 0.06282722513089005,
"grad_norm": 279.9869384765625,
"learning_rate": 3.020833333333333e-07,
"logits/chosen": -0.6136946678161621,
"logits/rejected": -0.605636477470398,
"logps/chosen": -278.4833679199219,
"logps/ref_chosen": -278.64752197265625,
"logps/ref_rejected": -249.309814453125,
"logps/rejected": -249.22158813476562,
"loss": 5.495,
"r_dpo/chosen_len": 270.8812561035156,
"r_dpo/length_delta": 25.228124618530273,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 245.6531219482422,
"step": 30
},
{
"epoch": 0.08376963350785341,
"grad_norm": 258.0906677246094,
"learning_rate": 4.0625e-07,
"logits/chosen": -0.6181729435920715,
"logits/rejected": -0.6429071426391602,
"logps/chosen": -282.7382507324219,
"logps/ref_chosen": -283.49981689453125,
"logps/ref_rejected": -265.32733154296875,
"logps/rejected": -265.03875732421875,
"loss": 5.3372,
"r_dpo/chosen_len": 281.43438720703125,
"r_dpo/length_delta": 33.34375,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 248.0906219482422,
"step": 40
},
{
"epoch": 0.10471204188481675,
"grad_norm": 267.5690002441406,
"learning_rate": 4.999932966293553e-07,
"logits/chosen": -0.6292977333068848,
"logits/rejected": -0.6698824763298035,
"logps/chosen": -278.5817565917969,
"logps/ref_chosen": -280.224365234375,
"logps/ref_rejected": -274.3541259765625,
"logps/rejected": -273.8267517089844,
"loss": 5.1553,
"r_dpo/chosen_len": 290.32501220703125,
"r_dpo/length_delta": 35.11249923706055,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 255.21249389648438,
"step": 50
},
{
"epoch": 0.1256544502617801,
"grad_norm": 258.2756652832031,
"learning_rate": 4.991893270335525e-07,
"logits/chosen": -0.6408439874649048,
"logits/rejected": -0.6542555093765259,
"logps/chosen": -278.864501953125,
"logps/ref_chosen": -281.12664794921875,
"logps/ref_rejected": -259.86456298828125,
"logps/rejected": -259.77728271484375,
"loss": 4.9283,
"r_dpo/chosen_len": 273.953125,
"r_dpo/length_delta": 29.084375381469727,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 244.86874389648438,
"step": 60
},
{
"epoch": 0.14659685863874344,
"grad_norm": 233.6326446533203,
"learning_rate": 4.970496218214204e-07,
"logits/chosen": -0.7015193700790405,
"logits/rejected": -0.7072200179100037,
"logps/chosen": -284.4013977050781,
"logps/ref_chosen": -287.71063232421875,
"logps/ref_rejected": -276.839599609375,
"logps/rejected": -276.4761047363281,
"loss": 4.76,
"r_dpo/chosen_len": 267.4937438964844,
"r_dpo/length_delta": 14.484375,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 253.00936889648438,
"step": 70
},
{
"epoch": 0.16753926701570682,
"grad_norm": 234.1904296875,
"learning_rate": 4.935856505068998e-07,
"logits/chosen": -0.6775287389755249,
"logits/rejected": -0.6742324829101562,
"logps/chosen": -275.77288818359375,
"logps/ref_chosen": -280.123046875,
"logps/ref_rejected": -258.8989562988281,
"logps/rejected": -258.0046691894531,
"loss": 4.6236,
"r_dpo/chosen_len": 267.4781188964844,
"r_dpo/length_delta": 32.46562576293945,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 235.0124969482422,
"step": 80
},
{
"epoch": 0.18848167539267016,
"grad_norm": 249.3949737548828,
"learning_rate": 4.8881598109976e-07,
"logits/chosen": -0.6876357197761536,
"logits/rejected": -0.694031834602356,
"logps/chosen": -273.43939208984375,
"logps/ref_chosen": -278.02545166015625,
"logps/ref_rejected": -251.0922393798828,
"logps/rejected": -251.4302978515625,
"loss": 4.5727,
"r_dpo/chosen_len": 274.20623779296875,
"r_dpo/length_delta": 44.97187423706055,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 229.234375,
"step": 90
},
{
"epoch": 0.2094240837696335,
"grad_norm": 247.84217834472656,
"learning_rate": 4.827661805750437e-07,
"logits/chosen": -0.6776241660118103,
"logits/rejected": -0.6956905126571655,
"logps/chosen": -269.8210754394531,
"logps/ref_chosen": -274.0089416503906,
"logps/ref_rejected": -274.14447021484375,
"logps/rejected": -275.0433349609375,
"loss": 4.6252,
"r_dpo/chosen_len": 275.3343811035156,
"r_dpo/length_delta": 21.912500381469727,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 253.421875,
"step": 100
},
{
"epoch": 0.23036649214659685,
"grad_norm": 227.6590576171875,
"learning_rate": 4.75468677825789e-07,
"logits/chosen": -0.701612114906311,
"logits/rejected": -0.7261066436767578,
"logps/chosen": -269.59173583984375,
"logps/ref_chosen": -273.23333740234375,
"logps/ref_rejected": -263.88787841796875,
"logps/rejected": -266.48272705078125,
"loss": 4.2996,
"r_dpo/chosen_len": 283.43438720703125,
"r_dpo/length_delta": 50.34375,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 233.0906219482422,
"step": 110
},
{
"epoch": 0.2513089005235602,
"grad_norm": 243.6549835205078,
"learning_rate": 4.669625898336438e-07,
"logits/chosen": -0.7124736905097961,
"logits/rejected": -0.7152305841445923,
"logps/chosen": -267.23974609375,
"logps/ref_chosen": -269.54656982421875,
"logps/ref_rejected": -272.7981872558594,
"logps/rejected": -275.8104553222656,
"loss": 4.5034,
"r_dpo/chosen_len": 264.7593688964844,
"r_dpo/length_delta": 13.840624809265137,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 250.9187469482422,
"step": 120
},
{
"epoch": 0.27225130890052357,
"grad_norm": 274.58111572265625,
"learning_rate": 4.5729351198915705e-07,
"logits/chosen": -0.7257631421089172,
"logits/rejected": -0.7154949903488159,
"logps/chosen": -272.1502990722656,
"logps/ref_chosen": -275.03448486328125,
"logps/ref_rejected": -276.39862060546875,
"logps/rejected": -279.466796875,
"loss": 4.445,
"r_dpo/chosen_len": 266.625,
"r_dpo/length_delta": 18.668750762939453,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 247.9562530517578,
"step": 130
},
{
"epoch": 0.2931937172774869,
"grad_norm": 247.52589416503906,
"learning_rate": 4.4651327368569684e-07,
"logits/chosen": -0.7425049543380737,
"logits/rejected": -0.7479813098907471,
"logps/chosen": -273.00714111328125,
"logps/ref_chosen": -276.0029602050781,
"logps/ref_rejected": -255.9320526123047,
"logps/rejected": -259.60784912109375,
"loss": 4.4958,
"r_dpo/chosen_len": 261.46875,
"r_dpo/length_delta": 22.375,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 239.09375,
"step": 140
},
{
"epoch": 0.31413612565445026,
"grad_norm": 264.3229675292969,
"learning_rate": 4.346796604970912e-07,
"logits/chosen": -0.7595945596694946,
"logits/rejected": -0.7660830616950989,
"logps/chosen": -294.779541015625,
"logps/ref_chosen": -298.2093505859375,
"logps/ref_rejected": -254.8907012939453,
"logps/rejected": -258.84454345703125,
"loss": 4.2489,
"r_dpo/chosen_len": 283.84375,
"r_dpo/length_delta": 48.359375,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 235.484375,
"step": 150
},
{
"epoch": 0.33507853403141363,
"grad_norm": 212.1317138671875,
"learning_rate": 4.218561044282098e-07,
"logits/chosen": -0.7484012246131897,
"logits/rejected": -0.7548493146896362,
"logps/chosen": -280.19281005859375,
"logps/ref_chosen": -281.94189453125,
"logps/ref_rejected": -255.5653533935547,
"logps/rejected": -261.9532470703125,
"loss": 4.0494,
"r_dpo/chosen_len": 267.828125,
"r_dpo/length_delta": 41.368751525878906,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 226.45938110351562,
"step": 160
},
{
"epoch": 0.35602094240837695,
"grad_norm": 263.6576843261719,
"learning_rate": 4.081113438988443e-07,
"logits/chosen": -0.7331063151359558,
"logits/rejected": -0.7357569932937622,
"logps/chosen": -287.6741638183594,
"logps/ref_chosen": -288.32232666015625,
"logps/ref_rejected": -239.85415649414062,
"logps/rejected": -246.4853973388672,
"loss": 4.1229,
"r_dpo/chosen_len": 285.203125,
"r_dpo/length_delta": 46.396873474121094,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 238.80624389648438,
"step": 170
},
{
"epoch": 0.3769633507853403,
"grad_norm": 204.8552703857422,
"learning_rate": 3.935190552834828e-07,
"logits/chosen": -0.7205232977867126,
"logits/rejected": -0.7394840121269226,
"logps/chosen": -285.03363037109375,
"logps/ref_chosen": -286.17889404296875,
"logps/ref_rejected": -249.9820098876953,
"logps/rejected": -256.336669921875,
"loss": 4.2693,
"r_dpo/chosen_len": 266.09063720703125,
"r_dpo/length_delta": 40.12812423706055,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 225.96249389648438,
"step": 180
},
{
"epoch": 0.39790575916230364,
"grad_norm": 227.1675567626953,
"learning_rate": 3.781574579820464e-07,
"logits/chosen": -0.7226096987724304,
"logits/rejected": -0.7422462701797485,
"logps/chosen": -279.1224365234375,
"logps/ref_chosen": -280.9278259277344,
"logps/ref_rejected": -254.3533477783203,
"logps/rejected": -262.07611083984375,
"loss": 4.0622,
"r_dpo/chosen_len": 276.33123779296875,
"r_dpo/length_delta": 41.993751525878906,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 234.33749389648438,
"step": 190
},
{
"epoch": 0.418848167539267,
"grad_norm": 257.41168212890625,
"learning_rate": 3.621088951385353e-07,
"logits/chosen": -0.741692841053009,
"logits/rejected": -0.7598770260810852,
"logps/chosen": -250.05776977539062,
"logps/ref_chosen": -253.1712188720703,
"logps/ref_rejected": -241.90478515625,
"logps/rejected": -246.9326629638672,
"loss": 4.178,
"r_dpo/chosen_len": 248.0749969482422,
"r_dpo/length_delta": 28.131250381469727,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 219.94375610351562,
"step": 200
},
{
"epoch": 0.418848167539267,
"eval_logits/chosen": -0.7478832006454468,
"eval_logits/rejected": -0.7490243911743164,
"eval_logps/chosen": -287.2392883300781,
"eval_logps/ref_chosen": -289.1346435546875,
"eval_logps/ref_rejected": -264.7782287597656,
"eval_logps/rejected": -271.315185546875,
"eval_loss": 0.5266098380088806,
"eval_r_dpo/chosen_len": 291.2619934082031,
"eval_r_dpo/length_delta": 42.86600112915039,
"eval_r_dpo/regularization_term": 0.0,
"eval_r_dpo/rejected_len": 248.39599609375,
"eval_runtime": 81.9324,
"eval_samples_per_second": 24.41,
"eval_steps_per_second": 1.526,
"step": 200
},
{
"epoch": 0.4397905759162304,
"grad_norm": 272.2992858886719,
"learning_rate": 3.454593922550693e-07,
"logits/chosen": -0.7123077511787415,
"logits/rejected": -0.7341128587722778,
"logps/chosen": -285.9898376464844,
"logps/ref_chosen": -287.9228210449219,
"logps/ref_rejected": -263.35595703125,
"logps/rejected": -270.154541015625,
"loss": 4.2219,
"r_dpo/chosen_len": 280.3125,
"r_dpo/length_delta": 36.68437576293945,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 243.6281280517578,
"step": 210
},
{
"epoch": 0.4607329842931937,
"grad_norm": 244.5552520751953,
"learning_rate": 3.2829819606729477e-07,
"logits/chosen": -0.7547723054885864,
"logits/rejected": -0.7528045773506165,
"logps/chosen": -280.8878479003906,
"logps/ref_chosen": -282.3331604003906,
"logps/ref_rejected": -272.5645446777344,
"logps/rejected": -279.7627868652344,
"loss": 4.0545,
"r_dpo/chosen_len": 261.359375,
"r_dpo/length_delta": 17.865625381469727,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 243.49374389648438,
"step": 220
},
{
"epoch": 0.4816753926701571,
"grad_norm": 290.22955322265625,
"learning_rate": 3.1071729615293424e-07,
"logits/chosen": -0.7329878211021423,
"logits/rejected": -0.7301102876663208,
"logps/chosen": -276.18194580078125,
"logps/ref_chosen": -276.1485595703125,
"logps/ref_rejected": -252.81198120117188,
"logps/rejected": -262.63323974609375,
"loss": 3.8953,
"r_dpo/chosen_len": 264.43438720703125,
"r_dpo/length_delta": 31.256250381469727,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 233.17813110351562,
"step": 230
},
{
"epoch": 0.5026178010471204,
"grad_norm": 242.89169311523438,
"learning_rate": 2.9281093183781403e-07,
"logits/chosen": -0.7076805233955383,
"logits/rejected": -0.729147732257843,
"logps/chosen": -270.7227783203125,
"logps/ref_chosen": -270.52520751953125,
"logps/ref_rejected": -254.83334350585938,
"logps/rejected": -263.5503234863281,
"loss": 4.1412,
"r_dpo/chosen_len": 271.81561279296875,
"r_dpo/length_delta": 37.099998474121094,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 234.7156219482422,
"step": 240
},
{
"epoch": 0.5235602094240838,
"grad_norm": 281.21990966796875,
"learning_rate": 2.7467508704251135e-07,
"logits/chosen": -0.7120658159255981,
"logits/rejected": -0.7247544527053833,
"logps/chosen": -288.3058166503906,
"logps/ref_chosen": -288.9503173828125,
"logps/ref_rejected": -265.0694580078125,
"logps/rejected": -273.70220947265625,
"loss": 4.2125,
"r_dpo/chosen_len": 277.50311279296875,
"r_dpo/length_delta": 41.103126525878906,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 236.39999389648438,
"step": 250
},
{
"epoch": 0.5445026178010471,
"grad_norm": 305.6956787109375,
"learning_rate": 2.5640697577740815e-07,
"logits/chosen": -0.7282508611679077,
"logits/rejected": -0.7263582944869995,
"logps/chosen": -287.1412353515625,
"logps/ref_chosen": -288.6393737792969,
"logps/ref_rejected": -265.315673828125,
"logps/rejected": -273.0039367675781,
"loss": 4.111,
"r_dpo/chosen_len": 271.48126220703125,
"r_dpo/length_delta": 24.390625,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 247.0906219482422,
"step": 260
},
{
"epoch": 0.5654450261780105,
"grad_norm": 248.96470642089844,
"learning_rate": 2.381045210440644e-07,
"logits/chosen": -0.7059388756752014,
"logits/rejected": -0.7142220735549927,
"logps/chosen": -279.7881774902344,
"logps/ref_chosen": -280.1373596191406,
"logps/ref_rejected": -264.84295654296875,
"logps/rejected": -273.27484130859375,
"loss": 4.011,
"r_dpo/chosen_len": 272.2875061035156,
"r_dpo/length_delta": 19.956249237060547,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 252.3312530517578,
"step": 270
},
{
"epoch": 0.5863874345549738,
"grad_norm": 216.84811401367188,
"learning_rate": 2.1986582993616925e-07,
"logits/chosen": -0.7371554374694824,
"logits/rejected": -0.7457467317581177,
"logps/chosen": -300.7189636230469,
"logps/ref_chosen": -301.8018798828125,
"logps/ref_rejected": -254.75112915039062,
"logps/rejected": -263.86175537109375,
"loss": 3.9179,
"r_dpo/chosen_len": 285.44061279296875,
"r_dpo/length_delta": 52.962501525878906,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 232.47811889648438,
"step": 280
},
{
"epoch": 0.6073298429319371,
"grad_norm": 258.0394592285156,
"learning_rate": 2.0178866775369774e-07,
"logits/chosen": -0.7233304977416992,
"logits/rejected": -0.7041932344436646,
"logps/chosen": -303.56158447265625,
"logps/ref_chosen": -302.79217529296875,
"logps/ref_rejected": -292.9220275878906,
"logps/rejected": -301.3799743652344,
"loss": 4.2767,
"r_dpo/chosen_len": 294.90625,
"r_dpo/length_delta": 20.774999618530273,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 274.1312561035156,
"step": 290
},
{
"epoch": 0.6282722513089005,
"grad_norm": 259.0774230957031,
"learning_rate": 1.839699339491937e-07,
"logits/chosen": -0.7432774305343628,
"logits/rejected": -0.7294503450393677,
"logps/chosen": -273.68316650390625,
"logps/ref_chosen": -275.8238220214844,
"logps/ref_rejected": -264.05743408203125,
"logps/rejected": -269.7857666015625,
"loss": 4.1741,
"r_dpo/chosen_len": 266.859375,
"r_dpo/length_delta": 20.734375,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 246.125,
"step": 300
},
{
"epoch": 0.6492146596858639,
"grad_norm": 330.3496398925781,
"learning_rate": 1.6650514271527465e-07,
"logits/chosen": -0.7020508646965027,
"logits/rejected": -0.7177096605300903,
"logps/chosen": -294.5872802734375,
"logps/ref_chosen": -296.6716003417969,
"logps/ref_rejected": -278.68426513671875,
"logps/rejected": -285.25396728515625,
"loss": 4.015,
"r_dpo/chosen_len": 292.91876220703125,
"r_dpo/length_delta": 32.55937576293945,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 260.359375,
"step": 310
},
{
"epoch": 0.6701570680628273,
"grad_norm": 258.3315734863281,
"learning_rate": 1.4948791099758052e-07,
"logits/chosen": -0.7137759923934937,
"logits/rejected": -0.7293487191200256,
"logps/chosen": -283.85919189453125,
"logps/ref_chosen": -284.1717529296875,
"logps/ref_rejected": -261.2606506347656,
"logps/rejected": -270.0918884277344,
"loss": 4.014,
"r_dpo/chosen_len": 279.90313720703125,
"r_dpo/length_delta": 44.537498474121094,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 235.36563110351562,
"step": 320
},
{
"epoch": 0.6910994764397905,
"grad_norm": 268.5577087402344,
"learning_rate": 1.3300945667758012e-07,
"logits/chosen": -0.7268847227096558,
"logits/rejected": -0.7360986471176147,
"logps/chosen": -284.22406005859375,
"logps/ref_chosen": -283.40338134765625,
"logps/ref_rejected": -271.27569580078125,
"logps/rejected": -280.9209899902344,
"loss": 4.0598,
"r_dpo/chosen_len": 267.67498779296875,
"r_dpo/length_delta": 13.015625,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 254.6593780517578,
"step": 330
},
{
"epoch": 0.7120418848167539,
"grad_norm": 247.73794555664062,
"learning_rate": 1.1715810961514072e-07,
"logits/chosen": -0.7288721203804016,
"logits/rejected": -0.7343642115592957,
"logps/chosen": -263.1542053222656,
"logps/ref_chosen": -259.82537841796875,
"logps/ref_rejected": -243.50222778320312,
"logps/rejected": -255.48574829101562,
"loss": 4.0065,
"r_dpo/chosen_len": 256.11248779296875,
"r_dpo/length_delta": 32.546875,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 223.5656280517578,
"step": 340
},
{
"epoch": 0.7329842931937173,
"grad_norm": 270.27557373046875,
"learning_rate": 1.0201883817182949e-07,
"logits/chosen": -0.7445515990257263,
"logits/rejected": -0.7549839615821838,
"logps/chosen": -299.8614501953125,
"logps/ref_chosen": -298.24725341796875,
"logps/ref_rejected": -272.657958984375,
"logps/rejected": -283.16290283203125,
"loss": 4.0568,
"r_dpo/chosen_len": 281.4624938964844,
"r_dpo/length_delta": 45.275001525878906,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 236.1875,
"step": 350
},
{
"epoch": 0.7539267015706806,
"grad_norm": 239.69332885742188,
"learning_rate": 8.76727937529367e-08,
"logits/chosen": -0.7207680344581604,
"logits/rejected": -0.7342051863670349,
"logps/chosen": -282.36883544921875,
"logps/ref_chosen": -281.881103515625,
"logps/ref_rejected": -265.4746398925781,
"logps/rejected": -275.1003112792969,
"loss": 4.1996,
"r_dpo/chosen_len": 272.64373779296875,
"r_dpo/length_delta": 30.071874618530273,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 242.57186889648438,
"step": 360
},
{
"epoch": 0.774869109947644,
"grad_norm": 198.85562133789062,
"learning_rate": 7.419687580962222e-08,
"logits/chosen": -0.7548901438713074,
"logits/rejected": -0.7558341026306152,
"logps/chosen": -302.01898193359375,
"logps/ref_chosen": -302.17822265625,
"logps/ref_rejected": -265.92877197265625,
"logps/rejected": -275.6526794433594,
"loss": 3.8979,
"r_dpo/chosen_len": 273.88751220703125,
"r_dpo/length_delta": 33.759376525878906,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 240.1281280517578,
"step": 370
},
{
"epoch": 0.7958115183246073,
"grad_norm": 286.9093933105469,
"learning_rate": 6.166331963291519e-08,
"logits/chosen": -0.7173658013343811,
"logits/rejected": -0.7228809595108032,
"logps/chosen": -300.5129699707031,
"logps/ref_chosen": -301.07965087890625,
"logps/ref_rejected": -266.7089538574219,
"logps/rejected": -274.7568359375,
"loss": 4.1149,
"r_dpo/chosen_len": 286.75311279296875,
"r_dpo/length_delta": 33.53125,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 253.2218780517578,
"step": 380
},
{
"epoch": 0.8167539267015707,
"grad_norm": 256.02996826171875,
"learning_rate": 5.013930914912476e-08,
"logits/chosen": -0.7370638251304626,
"logits/rejected": -0.7369734048843384,
"logps/chosen": -295.3925476074219,
"logps/ref_chosen": -296.6472473144531,
"logps/ref_rejected": -278.953857421875,
"logps/rejected": -287.25274658203125,
"loss": 4.0032,
"r_dpo/chosen_len": 287.91876220703125,
"r_dpo/length_delta": 30.081249237060547,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 257.8374938964844,
"step": 390
},
{
"epoch": 0.837696335078534,
"grad_norm": 229.88453674316406,
"learning_rate": 3.968661679220467e-08,
"logits/chosen": -0.726976752281189,
"logits/rejected": -0.7373378872871399,
"logps/chosen": -295.95574951171875,
"logps/ref_chosen": -296.6556091308594,
"logps/ref_rejected": -256.9266662597656,
"logps/rejected": -264.715576171875,
"loss": 4.0423,
"r_dpo/chosen_len": 278.96875,
"r_dpo/length_delta": 39.60625076293945,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 239.3625030517578,
"step": 400
},
{
"epoch": 0.837696335078534,
"eval_logits/chosen": -0.7442260980606079,
"eval_logits/rejected": -0.7460020780563354,
"eval_logps/chosen": -288.06787109375,
"eval_logps/ref_chosen": -289.1346435546875,
"eval_logps/ref_rejected": -264.7782287597656,
"eval_logps/rejected": -272.9751281738281,
"eval_loss": 0.5080092549324036,
"eval_r_dpo/chosen_len": 291.2619934082031,
"eval_r_dpo/length_delta": 42.86600112915039,
"eval_r_dpo/regularization_term": 0.0,
"eval_r_dpo/rejected_len": 248.39599609375,
"eval_runtime": 82.0059,
"eval_samples_per_second": 24.388,
"eval_steps_per_second": 1.524,
"step": 400
},
{
"epoch": 0.8586387434554974,
"grad_norm": 257.00933837890625,
"learning_rate": 3.036127238347164e-08,
"logits/chosen": -0.7183164358139038,
"logits/rejected": -0.7075697183609009,
"logps/chosen": -290.19342041015625,
"logps/ref_chosen": -290.08154296875,
"logps/ref_rejected": -272.56988525390625,
"logps/rejected": -280.45965576171875,
"loss": 4.0132,
"r_dpo/chosen_len": 282.40625,
"r_dpo/length_delta": 26.265625,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 256.140625,
"step": 410
},
{
"epoch": 0.8795811518324608,
"grad_norm": 258.52166748046875,
"learning_rate": 2.2213262793589482e-08,
"logits/chosen": -0.7189906239509583,
"logits/rejected": -0.7180750966072083,
"logps/chosen": -306.213623046875,
"logps/ref_chosen": -307.40240478515625,
"logps/ref_rejected": -279.85760498046875,
"logps/rejected": -287.7927551269531,
"loss": 3.9865,
"r_dpo/chosen_len": 296.8343811035156,
"r_dpo/length_delta": 37.165626525878906,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 259.66876220703125,
"step": 420
},
{
"epoch": 0.900523560209424,
"grad_norm": 225.66854858398438,
"learning_rate": 1.5286263996730026e-08,
"logits/chosen": -0.7201143503189087,
"logits/rejected": -0.7314451932907104,
"logps/chosen": -297.2125244140625,
"logps/ref_chosen": -297.7133483886719,
"logps/ref_rejected": -266.862060546875,
"logps/rejected": -276.1941223144531,
"loss": 4.0823,
"r_dpo/chosen_len": 290.703125,
"r_dpo/length_delta": 48.868751525878906,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 241.83438110351562,
"step": 430
},
{
"epoch": 0.9214659685863874,
"grad_norm": 296.7327880859375,
"learning_rate": 9.617406953185136e-09,
"logits/chosen": -0.7148677706718445,
"logits/rejected": -0.7244837880134583,
"logps/chosen": -293.4369201660156,
"logps/ref_chosen": -293.84521484375,
"logps/ref_rejected": -289.6698303222656,
"logps/rejected": -297.15863037109375,
"loss": 4.2486,
"r_dpo/chosen_len": 285.3656311035156,
"r_dpo/length_delta": 12.868749618530273,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 272.49688720703125,
"step": 440
},
{
"epoch": 0.9424083769633508,
"grad_norm": 255.14761352539062,
"learning_rate": 5.2370785753763356e-09,
"logits/chosen": -0.735411524772644,
"logits/rejected": -0.742519199848175,
"logps/chosen": -295.8509826660156,
"logps/ref_chosen": -297.1932373046875,
"logps/ref_rejected": -262.9091796875,
"logps/rejected": -271.23907470703125,
"loss": 3.8827,
"r_dpo/chosen_len": 282.81561279296875,
"r_dpo/length_delta": 40.631248474121094,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 242.1843719482422,
"step": 450
},
{
"epoch": 0.9633507853403142,
"grad_norm": 254.89694213867188,
"learning_rate": 2.168758844148272e-09,
"logits/chosen": -0.7343727946281433,
"logits/rejected": -0.7490428686141968,
"logps/chosen": -311.859619140625,
"logps/ref_chosen": -312.42291259765625,
"logps/ref_rejected": -278.7356262207031,
"logps/rejected": -286.7757263183594,
"loss": 4.1604,
"r_dpo/chosen_len": 288.9125061035156,
"r_dpo/length_delta": 43.23125076293945,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 245.68124389648438,
"step": 460
},
{
"epoch": 0.9842931937172775,
"grad_norm": 231.78506469726562,
"learning_rate": 4.288949484559934e-10,
"logits/chosen": -0.7186932563781738,
"logits/rejected": -0.7156038284301758,
"logps/chosen": -276.20965576171875,
"logps/ref_chosen": -278.13385009765625,
"logps/ref_rejected": -256.70648193359375,
"logps/rejected": -264.2013244628906,
"loss": 3.9555,
"r_dpo/chosen_len": 268.8687438964844,
"r_dpo/length_delta": 26.262500762939453,
"r_dpo/regularization_term": 0.0,
"r_dpo/rejected_len": 242.6062469482422,
"step": 470
},
{
"epoch": 0.9989528795811519,
"step": 477,
"total_flos": 0.0,
"train_loss": 4.32008807264284,
"train_runtime": 6099.6278,
"train_samples_per_second": 10.023,
"train_steps_per_second": 0.078
}
],
"logging_steps": 10,
"max_steps": 477,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 50,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": false,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}