Adding Evaluation Results (#7)

- Adding Evaluation Results (796f0cfbe24fe87fbde205993ded00360f817e5e)

Co-authored-by: Open LLM Leaderboard PR Bot <leaderboard-pr-bot@users.noreply.huggingface.co>
This commit is contained in:
ai-modelscope
2025-01-03 12:05:54 +08:00
parent a5e91d812f
commit b61537282f
36 changed files with 93873 additions and 65 deletions

16
eval_results.json Normal file
View File

@@ -0,0 +1,16 @@
{
"epoch": 1.0,
"eval_logits/chosen": -1.2132024765014648,
"eval_logits/rejected": -1.2924575805664062,
"eval_logps/chosen": -396.52020263671875,
"eval_logps/rejected": -409.31964111328125,
"eval_loss": 0.4216844439506531,
"eval_rewards/accuracies": 0.8134920597076416,
"eval_rewards/chosen": -0.1933162659406662,
"eval_rewards/margins": 2.1035311222076416,
"eval_rewards/rejected": -2.296847343444824,
"eval_runtime": 398.9749,
"eval_samples": 2000,
"eval_samples_per_second": 5.013,
"eval_steps_per_second": 0.158
}