Files
QWEN3-4B-CPT/eval/eval_results_final.json

15657 lines
521 KiB
JSON
Raw Permalink Normal View History

{
"model_path": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"ppl": null,
"base_ppl": null,
"benchmarks": {
"cpt": {
"results": {
"arc_challenge": {
"alias": "arc_challenge",
"acc,none": 0.4825,
"acc_stderr,none": 0.025015972341295333,
"acc_norm,none": 0.5325,
"acc_norm_stderr,none": 0.024978374105060028
},
"arc_easy": {
"alias": "arc_easy",
"acc,none": 0.78,
"acc_stderr,none": 0.020738254217024313,
"acc_norm,none": 0.795,
"acc_norm_stderr,none": 0.020210359883399975
},
"hellaswag": {
"alias": "hellaswag",
"acc,none": 0.4975,
"acc_stderr,none": 0.025030995822773405,
"acc_norm,none": 0.63,
"acc_norm_stderr,none": 0.024170447375168467
},
"kmmlu": {
"acc,none": 0.4692806221646144,
"acc_stderr,none": 0.0039182515413587,
"alias": "kmmlu"
},
"kmmlu_applied_science": {
"acc,none": 0.45375,
"acc_stderr,none": 0.007111885914543827,
"alias": " - kmmlu_applied_science"
},
"kmmlu_aviation_engineering_and_maintenance": {
"alias": " - kmmlu_aviation_engineering_and_maintenance",
"acc,none": 0.46,
"acc_stderr,none": 0.024951079956135092
},
"kmmlu_electronics_engineering": {
"alias": " - kmmlu_electronics_engineering",
"acc,none": 0.6275,
"acc_stderr,none": 0.0242038000082031
},
"kmmlu_energy_management": {
"alias": " - kmmlu_energy_management",
"acc,none": 0.395,
"acc_stderr,none": 0.0244731452227279
},
"kmmlu_environmental_science": {
"alias": " - kmmlu_environmental_science",
"acc,none": 0.37,
"acc_stderr,none": 0.024170447375168453
},
"kmmlu_gas_technology_and_engineering": {
"alias": " - kmmlu_gas_technology_and_engineering",
"acc,none": 0.405,
"acc_stderr,none": 0.024575340657273674
},
"kmmlu_geomatics": {
"alias": " - kmmlu_geomatics",
"acc,none": 0.425,
"acc_stderr,none": 0.024748104405776187
},
"kmmlu_industrial_engineer": {
"alias": " - kmmlu_industrial_engineer",
"acc,none": 0.4275,
"acc_stderr,none": 0.024766769210836766
},
"kmmlu_machine_design_and_manufacturing": {
"alias": " - kmmlu_machine_design_and_manufacturing",
"acc,none": 0.4975,
"acc_stderr,none": 0.025030995822773395
},
"kmmlu_maritime_engineering": {
"alias": " - kmmlu_maritime_engineering",
"acc,none": 0.4075,
"acc_stderr,none": 0.02459923129797198
},
"kmmlu_nondestructive_testing": {
"alias": " - kmmlu_nondestructive_testing",
"acc,none": 0.475,
"acc_stderr,none": 0.024999999999999994
},
"kmmlu_railway_and_automotive_engineering": {
"alias": " - kmmlu_railway_and_automotive_engineering",
"acc,none": 0.3825,
"acc_stderr,none": 0.024330316186072946
},
"kmmlu_telecommunications_and_wireless_technology": {
"alias": " - kmmlu_telecommunications_and_wireless_technology",
"acc,none": 0.5725,
"acc_stderr,none": 0.02476676921083677
},
"kmmlu_humss": {
"acc,none": 0.4776556776556777,
"acc_stderr,none": 0.00943997794327789,
"alias": " - kmmlu_humss"
},
"kmmlu_accounting": {
"alias": " - kmmlu_accounting",
"acc,none": 0.5,
"acc_stderr,none": 0.050251890762960605
},
"kmmlu_criminal_law": {
"alias": " - kmmlu_criminal_law",
"acc,none": 0.39,
"acc_stderr,none": 0.03457567623250012
},
"kmmlu_economics": {
"alias": " - kmmlu_economics",
"acc,none": 0.5461538461538461,
"acc_stderr,none": 0.04383459241436368
},
"kmmlu_education": {
"alias": " - kmmlu_education",
"acc,none": 0.64,
"acc_stderr,none": 0.048241815132442176
},
"kmmlu_korean_history": {
"alias": " - kmmlu_korean_history",
"acc,none": 0.3,
"acc_stderr,none": 0.046056618647183814
},
"kmmlu_law": {
"alias": " - kmmlu_law",
"acc,none": 0.375,
"acc_stderr,none": 0.02423646044779629
},
"kmmlu_management": {
"alias": " - kmmlu_management",
"acc,none": 0.5225,
"acc_stderr,none": 0.02500595167250431
},
"kmmlu_political_science_and_sociology": {
"alias": " - kmmlu_political_science_and_sociology",
"acc,none": 0.55,
"acc_stderr,none": 0.02877080459987894
},
"kmmlu_psychology": {
"alias": " - kmmlu_psychology",
"acc,none": 0.45,
"acc_stderr,none": 0.024905837706844923
},
"kmmlu_social_welfare": {
"alias": " - kmmlu_social_welfare",
"acc,none": 0.57,
"acc_stderr,none": 0.02478478796128207
},
"kmmlu_taxation": {
"alias": " - kmmlu_taxation",
"acc,none": 0.395,
"acc_stderr,none": 0.03465370682892271
},
"kmmlu_other": {
"acc,none": 0.4697222222222222,
"acc_stderr,none": 0.008043980393376315,
"alias": " - kmmlu_other"
},
"kmmlu_agricultural_sciences": {
"alias": " - kmmlu_agricultural_sciences",
"acc,none": 0.3625,
"acc_stderr,none": 0.024066207238097735
},
"kmmlu_construction": {
"alias": " - kmmlu_construction",
"acc,none": 0.4,
"acc_stderr,none": 0.024525573579398552
},
"kmmlu_fashion": {
"alias": " - kmmlu_fashion",
"acc,none": 0.45,
"acc_stderr,none": 0.024905837706844923
},
"kmmlu_food_processing": {
"alias": " - kmmlu_food_processing",
"acc,none": 0.3675,
"acc_stderr,none": 0.024136399679191744
},
"kmmlu_health": {
"alias": " - kmmlu_health",
"acc,none": 0.58,
"acc_stderr,none": 0.049604496374885836
},
"kmmlu_interior_architecture_and_design": {
"alias": " - kmmlu_interior_architecture_and_design",
"acc,none": 0.6175,
"acc_stderr,none": 0.024330316186072936
},
"kmmlu_marketing": {
"alias": " - kmmlu_marketing",
"acc,none": 0.765,
"acc_stderr,none": 0.021226490755055
},
"kmmlu_patent": {
"alias": " - kmmlu_patent",
"acc,none": 0.42,
"acc_stderr,none": 0.049604496374885836
},
"kmmlu_public_safety": {
"alias": " - kmmlu_public_safety",
"acc,none": 0.38,
"acc_stderr,none": 0.024299715851758236
},
"kmmlu_real_estate": {
"alias": " - kmmlu_real_estate",
"acc,none": 0.45,
"acc_stderr,none": 0.03526639466921485
},
"kmmlu_refrigerating_machinery": {
"alias": " - kmmlu_refrigerating_machinery",
"acc,none": 0.41,
"acc_stderr,none": 0.02462246259333947
},
"kmmlu_stem": {
"acc,none": 0.48093023255813955,
"acc_stderr,none": 0.007306868046626305,
"alias": " - kmmlu_stem"
},
"kmmlu_biology": {
"alias": " - kmmlu_biology",
"acc,none": 0.3125,
"acc_stderr,none": 0.023204644228784484
},
"kmmlu_chemical_engineering": {
"alias": " - kmmlu_chemical_engineering",
"acc,none": 0.4875,
"acc_stderr,none": 0.025023485209500245
},
"kmmlu_chemistry": {
"alias": " - kmmlu_chemistry",
"acc,none": 0.5175,
"acc_stderr,none": 0.025015972341295323
},
"kmmlu_civil_engineering": {
"alias": " - kmmlu_civil_engineering",
"acc,none": 0.3925,
"acc_stderr,none": 0.024445927747963322
},
"kmmlu_computer_science": {
"alias": " - kmmlu_computer_science",
"acc,none": 0.74,
"acc_stderr,none": 0.021959178349484305
},
"kmmlu_ecology": {
"alias": " - kmmlu_ecology",
"acc,none": 0.505,
"acc_stderr,none": 0.02503005711936146
},
"kmmlu_electrical_engineering": {
"alias": " - kmmlu_electrical_engineering",
"acc,none": 0.3425,
"acc_stderr,none": 0.02375700661717548
},
"kmmlu_information_technology": {
"alias": " - kmmlu_information_technology",
"acc,none": 0.7525,
"acc_stderr,none": 0.021605006729678956
},
"kmmlu_materials_engineering": {
"alias": " - kmmlu_materials_engineering",
"acc,none": 0.475,
"acc_stderr,none": 0.025
},
"kmmlu_math": {
"alias": " - kmmlu_math",
"acc,none": 0.3333333333333333,
"acc_stderr,none": 0.027262027336984393
},
"kmmlu_mechanical_engineering": {
"alias": " - kmmlu_mechanical_engineering",
"acc,none": 0.395,
"acc_stderr,none": 0.0244731452227279
},
"kobest_boolq": {
"alias": "kobest_boolq",
"acc,none": 0.755,
"acc_stderr,none": 0.02153129097913247,
"f1,none": 0.7379609080456697,
"f1_stderr,none": "N/A"
},
"kobest_copa": {
"alias": "kobest_copa",
"acc,none": 0.6525,
"acc_stderr,none": 0.023838625698390636,
"f1,none": 0.6523935455233165,
"f1_stderr,none": "N/A"
},
"kobest_hellaswag": {
"alias": "kobest_hellaswag",
"acc,none": 0.4325,
"acc_stderr,none": 0.024802162065186355,
"f1,none": 0.4264529493583016,
"f1_stderr,none": "N/A",
"acc_norm,none": 0.565,
"acc_norm_stderr,none": 0.024818892876375884
},
"mmlu": {
"acc,none": 0.7352865587252634,
"acc_stderr,none": 0.003887849176172822,
"alias": "mmlu"
},
"mmlu_humanities": {
"acc,none": 0.6862808842652796,
"acc_stderr,none": 0.0077616777391173045,
"alias": " - humanities"
},
"mmlu_formal_logic": {
"alias": " - formal_logic",
"acc,none": 0.5873015873015873,
"acc_stderr,none": 0.04403438954768177
},
"mmlu_high_school_european_history": {
"alias": " - high_school_european_history",
"acc,none": 0.7818181818181819,
"acc_stderr,none": 0.03225078108306289
},
"mmlu_high_school_us_history": {
"alias": " - high_school_us_history",
"acc,none": 0.8186274509803921,
"acc_stderr,none": 0.02704462171947408
},
"mmlu_high_school_world_history": {
"alias": " - high_school_world_history",
"acc,none": 0.8481012658227848,
"acc_stderr,none": 0.023363878096632453
},
"mmlu_international_law": {
"alias": " - international_law",
"acc,none": 0.8264462809917356,
"acc_stderr,none": 0.0345727283691767
},
"mmlu_jurisprudence": {
"alias": " - jurisprudence",
"acc,none": 0.8148148148148148,
"acc_stderr,none": 0.03755265865037183
},
"mmlu_logical_fallacies": {
"alias": " - logical_fallacies",
"acc,none": 0.8466257668711656,
"acc_stderr,none": 0.02831160144143859
},
"mmlu_moral_disputes": {
"alias": " - moral_disputes",
"acc,none": 0.7543352601156069,
"acc_stderr,none": 0.023176298203992005
},
"mmlu_moral_scenarios": {
"alias": " - moral_scenarios",
"acc,none": 0.3225,
"acc_stderr,none": 0.023400926978618716
},
"mmlu_philosophy": {
"alias": " - philosophy",
"acc,none": 0.7331189710610932,
"acc_stderr,none": 0.025122637608816636
},
"mmlu_prehistory": {
"alias": " - prehistory",
"acc,none": 0.7870370370370371,
"acc_stderr,none": 0.02277971908873339
},
"mmlu_professional_law": {
"alias": " - professional_law",
"acc,none": 0.5075,
"acc_stderr,none": 0.02502849253543831
},
"mmlu_world_religions": {
"alias": " - world_religions",
"acc,none": 0.8070175438596491,
"acc_stderr,none": 0.030267457554898458
},
"mmlu_other": {
"acc,none": 0.7415565345080763,
"acc_stderr,none": 0.008104267812218218,
"alias": " - other"
},
"mmlu_business_ethics": {
"alias": " - business_ethics",
"acc,none": 0.76,
"acc_stderr,none": 0.04292346959909282
},
"mmlu_clinical_knowledge": {
"alias": " - clinical_knowledge",
"acc,none": 0.769811320754717,
"acc_stderr,none": 0.025907897122408173
},
"mmlu_college_medicine": {
"alias": " - college_medicine",
"acc,none": 0.7456647398843931,
"acc_stderr,none": 0.0332055644308557
},
"mmlu_global_facts": {
"alias": " - global_facts",
"acc,none": 0.44,
"acc_stderr,none": 0.0498887651569859
},
"mmlu_human_aging": {
"alias": " - human_aging",
"acc,none": 0.7399103139013453,
"acc_stderr,none": 0.029442495585857473
},
"mmlu_management": {
"alias": " - management",
"acc,none": 0.8640776699029126,
"acc_stderr,none": 0.0339329572976101
},
"mmlu_marketing": {
"alias": " - marketing",
"acc,none": 0.8931623931623932,
"acc_stderr,none": 0.020237149008990932
},
"mmlu_medical_genetics": {
"alias": " - medical_genetics",
"acc,none": 0.8,
"acc_stderr,none": 0.04020151261036846
},
"mmlu_miscellaneous": {
"alias": " - miscellaneous",
"acc,none": 0.8225,
"acc_stderr,none": 0.019128489820344343
},
"mmlu_nutrition": {
"alias": " - nutrition",
"acc,none": 0.7777777777777778,
"acc_stderr,none": 0.02380518652488816
},
"mmlu_professional_accounting": {
"alias": " - professional_accounting",
"acc,none": 0.574468085106383,
"acc_stderr,none": 0.029494827600144366
},
"mmlu_professional_medicine": {
"alias": " - professional_medicine",
"acc,none": 0.7757352941176471,
"acc_stderr,none": 0.02533684856333236
},
"mmlu_virology": {
"alias": " - virology",
"acc,none": 0.5060240963855421,
"acc_stderr,none": 0.038922121953330446
},
"mmlu_social_sciences": {
"acc,none": 0.8158088235294118,
"acc_stderr,none": 0.007306038192044323,
"alias": " - social sciences"
},
"mmlu_econometrics": {
"alias": " - econometrics",
"acc,none": 0.6578947368421053,
"acc_stderr,none": 0.04462917535336937
},
"mmlu_high_school_geography": {
"alias": " - high_school_geography",
"acc,none": 0.8585858585858586,
"acc_stderr,none": 0.02482590979334335
},
"mmlu_high_school_government_and_politics": {
"alias": " - high_school_government_and_politics",
"acc,none": 0.8704663212435233,
"acc_stderr,none": 0.024233532297758716
},
"mmlu_high_school_macroeconomics": {
"alias": " - high_school_macroeconomics",
"acc,none": 0.8076923076923077,
"acc_stderr,none": 0.019982347208637296
},
"mmlu_high_school_microeconomics": {
"alias": " - high_school_microeconomics",
"acc,none": 0.8991596638655462,
"acc_stderr,none": 0.019559663430480802
},
"mmlu_high_school_psychology": {
"alias": " - high_school_psychology",
"acc,none": 0.905,
"acc_stderr,none": 0.014679107277903242
},
"mmlu_human_sexuality": {
"alias": " - human_sexuality",
"acc,none": 0.7786259541984732,
"acc_stderr,none": 0.03641297081313729
},
"mmlu_professional_psychology": {
"alias": " - professional_psychology",
"acc,none": 0.74,
"acc_stderr,none": 0.02195917834948431
},
"mmlu_public_relations": {
"alias": " - public_relations",
"acc,none": 0.6727272727272727,
"acc_stderr,none": 0.0449429086625209
},
"mmlu_security_studies": {
"alias": " - security_studies",
"acc,none": 0.7428571428571429,
"acc_stderr,none": 0.027979823538744546
},
"mmlu_sociology": {
"alias": " - sociology",
"acc,none": 0.8557213930348259,
"acc_stderr,none": 0.02484575321230605
},
"mmlu_us_foreign_policy": {
"alias": " - us_foreign_policy",
"acc,none": 0.89,
"acc_stderr,none": 0.03144660377352203
},
"mmlu_stem": {
"acc,none": 0.7082143989850935,
"acc_stderr,none": 0.007816574368205405,
"alias": " - stem"
},
"mmlu_abstract_algebra": {
"alias": " - abstract_algebra",
"acc,none": 0.46,
"acc_stderr,none": 0.05009082659620333
},
"mmlu_anatomy": {
"alias": " - anatomy",
"acc,none": 0.7111111111111111,
"acc_stderr,none": 0.0391545063041425
},
"mmlu_astronomy": {
"alias": " - astronomy",
"acc,none": 0.8486842105263158,
"acc_stderr,none": 0.029162631596843975
},
"mmlu_college_biology": {
"alias": " - college_biology",
"acc,none": 0.8263888888888888,
"acc_stderr,none": 0.03167473383795717
},
"mmlu_college_chemistry": {
"alias": " - college_chemistry",
"acc,none": 0.52,
"acc_stderr,none": 0.050211673156867795
},
"mmlu_college_computer_science": {
"alias": " - college_computer_science",
"acc,none": 0.68,
"acc_stderr,none": 0.04688261722621504
},
"mmlu_college_mathematics": {
"alias": " - college_mathematics",
"acc,none": 0.53,
"acc_stderr,none": 0.05016135580465919
},
"mmlu_college_physics": {
"alias": " - college_physics",
"acc,none": 0.5784313725490197,
"acc_stderr,none": 0.049135952012745045
},
"mmlu_computer_security": {
"alias": " - computer_security",
"acc,none": 0.83,
"acc_stderr,none": 0.03775251680686371
},
"mmlu_conceptual_physics": {
"alias": " - conceptual_physics",
"acc,none": 0.8,
"acc_stderr,none": 0.026148818018424506
},
"mmlu_electrical_engineering": {
"alias": " - electrical_engineering",
"acc,none": 0.7586206896551724,
"acc_stderr,none": 0.03565998174135302
},
"mmlu_elementary_mathematics": {
"alias": " - elementary_mathematics",
"acc,none": 0.6746031746031746,
"acc_stderr,none": 0.024130158299762613
},
"mmlu_high_school_biology": {
"alias": " - high_school_biology",
"acc,none": 0.9,
"acc_stderr,none": 0.017066403719657258
},
"mmlu_high_school_chemistry": {
"alias": " - high_school_chemistry",
"acc,none": 0.729064039408867,
"acc_stderr,none": 0.03127090713297698
},
"mmlu_high_school_computer_science": {
"alias": " - high_school_computer_science",
"acc,none": 0.85,
"acc_stderr,none": 0.0358870281282637
},
"mmlu_high_school_mathematics": {
"alias": " - high_school_mathematics",
"acc,none": 0.5296296296296297,
"acc_stderr,none": 0.030431963547936584
},
"mmlu_high_school_physics": {
"alias": " - high_school_physics",
"acc,none": 0.6754966887417219,
"acc_stderr,none": 0.03822746937658752
},
"mmlu_high_school_statistics": {
"alias": " - high_school_statistics",
"acc,none": 0.7037037037037037,
"acc_stderr,none": 0.031141447823536044
},
"mmlu_machine_learning": {
"alias": " - machine_learning",
"acc,none": 0.5892857142857143,
"acc_stderr,none": 0.04669510663875191
},
"winogrande": {
"alias": "winogrande",
"acc,none": 0.7225,
"acc_stderr,none": 0.022416302137144652
}
},
"groups": {
"kmmlu": {
"acc,none": 0.4692806221646144,
"acc_stderr,none": 0.0039182515413587,
"alias": "kmmlu"
},
"kmmlu_applied_science": {
"acc,none": 0.45375,
"acc_stderr,none": 0.007111885914543827,
"alias": " - kmmlu_applied_science"
},
"kmmlu_humss": {
"acc,none": 0.4776556776556777,
"acc_stderr,none": 0.00943997794327789,
"alias": " - kmmlu_humss"
},
"kmmlu_other": {
"acc,none": 0.4697222222222222,
"acc_stderr,none": 0.008043980393376315,
"alias": " - kmmlu_other"
},
"kmmlu_stem": {
"acc,none": 0.48093023255813955,
"acc_stderr,none": 0.007306868046626305,
"alias": " - kmmlu_stem"
},
"mmlu": {
"acc,none": 0.7352865587252634,
"acc_stderr,none": 0.003887849176172822,
"alias": "mmlu"
},
"mmlu_humanities": {
"acc,none": 0.6862808842652796,
"acc_stderr,none": 0.0077616777391173045,
"alias": " - humanities"
},
"mmlu_other": {
"acc,none": 0.7415565345080763,
"acc_stderr,none": 0.008104267812218218,
"alias": " - other"
},
"mmlu_social_sciences": {
"acc,none": 0.8158088235294118,
"acc_stderr,none": 0.007306038192044323,
"alias": " - social sciences"
},
"mmlu_stem": {
"acc,none": 0.7082143989850935,
"acc_stderr,none": 0.007816574368205405,
"alias": " - stem"
}
},
"group_subtasks": {
"mmlu_humanities": [
"mmlu_formal_logic",
"mmlu_high_school_european_history",
"mmlu_high_school_us_history",
"mmlu_high_school_world_history",
"mmlu_international_law",
"mmlu_jurisprudence",
"mmlu_logical_fallacies",
"mmlu_moral_disputes",
"mmlu_moral_scenarios",
"mmlu_philosophy",
"mmlu_prehistory",
"mmlu_professional_law",
"mmlu_world_religions"
],
"mmlu_social_sciences": [
"mmlu_econometrics",
"mmlu_high_school_geography",
"mmlu_high_school_government_and_politics",
"mmlu_high_school_macroeconomics",
"mmlu_high_school_microeconomics",
"mmlu_high_school_psychology",
"mmlu_human_sexuality",
"mmlu_professional_psychology",
"mmlu_public_relations",
"mmlu_security_studies",
"mmlu_sociology",
"mmlu_us_foreign_policy"
],
"mmlu_other": [
"mmlu_business_ethics",
"mmlu_clinical_knowledge",
"mmlu_college_medicine",
"mmlu_global_facts",
"mmlu_human_aging",
"mmlu_management",
"mmlu_marketing",
"mmlu_medical_genetics",
"mmlu_miscellaneous",
"mmlu_nutrition",
"mmlu_professional_accounting",
"mmlu_professional_medicine",
"mmlu_virology"
],
"mmlu_stem": [
"mmlu_abstract_algebra",
"mmlu_anatomy",
"mmlu_astronomy",
"mmlu_college_biology",
"mmlu_college_chemistry",
"mmlu_college_computer_science",
"mmlu_college_mathematics",
"mmlu_college_physics",
"mmlu_computer_security",
"mmlu_conceptual_physics",
"mmlu_electrical_engineering",
"mmlu_elementary_mathematics",
"mmlu_high_school_biology",
"mmlu_high_school_chemistry",
"mmlu_high_school_computer_science",
"mmlu_high_school_mathematics",
"mmlu_high_school_physics",
"mmlu_high_school_statistics",
"mmlu_machine_learning"
],
"mmlu": [
"mmlu_stem",
"mmlu_other",
"mmlu_social_sciences",
"mmlu_humanities"
],
"hellaswag": [],
"arc_easy": [],
"arc_challenge": [],
"winogrande": [],
"kmmlu_humss": [
"kmmlu_accounting",
"kmmlu_criminal_law",
"kmmlu_economics",
"kmmlu_education",
"kmmlu_korean_history",
"kmmlu_law",
"kmmlu_management",
"kmmlu_political_science_and_sociology",
"kmmlu_psychology",
"kmmlu_social_welfare",
"kmmlu_taxation"
],
"kmmlu_applied_science": [
"kmmlu_aviation_engineering_and_maintenance",
"kmmlu_electronics_engineering",
"kmmlu_energy_management",
"kmmlu_environmental_science",
"kmmlu_gas_technology_and_engineering",
"kmmlu_geomatics",
"kmmlu_industrial_engineer",
"kmmlu_machine_design_and_manufacturing",
"kmmlu_maritime_engineering",
"kmmlu_nondestructive_testing",
"kmmlu_railway_and_automotive_engineering",
"kmmlu_telecommunications_and_wireless_technology"
],
"kmmlu_other": [
"kmmlu_agricultural_sciences",
"kmmlu_construction",
"kmmlu_fashion",
"kmmlu_food_processing",
"kmmlu_health",
"kmmlu_interior_architecture_and_design",
"kmmlu_marketing",
"kmmlu_patent",
"kmmlu_public_safety",
"kmmlu_real_estate",
"kmmlu_refrigerating_machinery"
],
"kmmlu_stem": [
"kmmlu_biology",
"kmmlu_chemical_engineering",
"kmmlu_chemistry",
"kmmlu_civil_engineering",
"kmmlu_computer_science",
"kmmlu_ecology",
"kmmlu_electrical_engineering",
"kmmlu_information_technology",
"kmmlu_materials_engineering",
"kmmlu_math",
"kmmlu_mechanical_engineering"
],
"kmmlu": [
"kmmlu_stem",
"kmmlu_other",
"kmmlu_applied_science",
"kmmlu_humss"
],
"kobest_boolq": [],
"kobest_copa": [],
"kobest_hellaswag": []
},
"configs": {
"arc_challenge": {
"task": "arc_challenge",
"tag": [
"ai2_arc"
],
"dataset_path": "allenai/ai2_arc",
"dataset_name": "ARC-Challenge",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": "{{choices.label.index(answerKey)}}",
"unsafe_code": false,
"doc_to_choice": "{{choices.text}}",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": null,
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_choice": "{{choices.text}}",
"doc_to_target": "{{choices.label.index(answerKey)}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "acc_norm",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": true,
"doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"arc_easy": {
"task": "arc_easy",
"tag": [
"ai2_arc"
],
"dataset_path": "allenai/ai2_arc",
"dataset_name": "ARC-Easy",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": "{{choices.label.index(answerKey)}}",
"unsafe_code": false,
"doc_to_choice": "{{choices.text}}",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": null,
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_choice": "{{choices.text}}",
"doc_to_target": "{{choices.label.index(answerKey)}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "acc_norm",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": true,
"doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"hellaswag": {
"task": "hellaswag",
"tag": [
"multiple_choice"
],
"dataset_path": "Rowan/hellaswag",
"training_split": "train",
"validation_split": "validation",
"process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc):\n ctx = doc[\"ctx_a\"] + \" \" + doc[\"ctx_b\"].capitalize()\n out_doc = {\n \"query\": preprocess(doc[\"activity_label\"] + \": \" + ctx),\n \"choices\": [preprocess(ending) for ending in doc[\"endings\"]],\n \"gold\": int(doc[\"label\"]),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
"doc_to_text": "{{query}}",
"doc_to_target": "{{label}}",
"unsafe_code": false,
"doc_to_choice": "choices",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": null,
"process_docs": "<function process_docs at 0x77cd4ebe45e0>",
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{query}}",
"doc_to_choice": "choices",
"doc_to_target": "{{label}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "acc_norm",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_accounting": {
"task": "kmmlu_accounting",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Accounting",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_agricultural_sciences": {
"task": "kmmlu_agricultural_sciences",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Agricultural-Sciences",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_aviation_engineering_and_maintenance": {
"task": "kmmlu_aviation_engineering_and_maintenance",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Aviation-Engineering-and-Maintenance",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_biology": {
"task": "kmmlu_biology",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Biology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_chemical_engineering": {
"task": "kmmlu_chemical_engineering",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Chemical-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_chemistry": {
"task": "kmmlu_chemistry",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Chemistry",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_civil_engineering": {
"task": "kmmlu_civil_engineering",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Civil-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_computer_science": {
"task": "kmmlu_computer_science",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Computer-Science",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_construction": {
"task": "kmmlu_construction",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Construction",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_criminal_law": {
"task": "kmmlu_criminal_law",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Criminal-Law",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_ecology": {
"task": "kmmlu_ecology",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Ecology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_economics": {
"task": "kmmlu_economics",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Economics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_education": {
"task": "kmmlu_education",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Education",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_electrical_engineering": {
"task": "kmmlu_electrical_engineering",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Electrical-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_electronics_engineering": {
"task": "kmmlu_electronics_engineering",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Electronics-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_energy_management": {
"task": "kmmlu_energy_management",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Energy-Management",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_environmental_science": {
"task": "kmmlu_environmental_science",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Environmental-Science",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_fashion": {
"task": "kmmlu_fashion",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Fashion",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_food_processing": {
"task": "kmmlu_food_processing",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Food-Processing",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_gas_technology_and_engineering": {
"task": "kmmlu_gas_technology_and_engineering",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Gas-Technology-and-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_geomatics": {
"task": "kmmlu_geomatics",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Geomatics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_health": {
"task": "kmmlu_health",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Health",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_industrial_engineer": {
"task": "kmmlu_industrial_engineer",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Industrial-Engineer",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_information_technology": {
"task": "kmmlu_information_technology",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Information-Technology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_interior_architecture_and_design": {
"task": "kmmlu_interior_architecture_and_design",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Interior-Architecture-and-Design",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_korean_history": {
"task": "kmmlu_korean_history",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Korean-History",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_law": {
"task": "kmmlu_law",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Law",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_machine_design_and_manufacturing": {
"task": "kmmlu_machine_design_and_manufacturing",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Machine-Design-and-Manufacturing",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_management": {
"task": "kmmlu_management",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Management",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_maritime_engineering": {
"task": "kmmlu_maritime_engineering",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Maritime-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_marketing": {
"task": "kmmlu_marketing",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Marketing",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_materials_engineering": {
"task": "kmmlu_materials_engineering",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Materials-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_math": {
"task": "kmmlu_math",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Math",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_mechanical_engineering": {
"task": "kmmlu_mechanical_engineering",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Mechanical-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_nondestructive_testing": {
"task": "kmmlu_nondestructive_testing",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Nondestructive-Testing",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_patent": {
"task": "kmmlu_patent",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Patent",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_political_science_and_sociology": {
"task": "kmmlu_political_science_and_sociology",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Political-Science-and-Sociology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_psychology": {
"task": "kmmlu_psychology",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Psychology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_public_safety": {
"task": "kmmlu_public_safety",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Public-Safety",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_railway_and_automotive_engineering": {
"task": "kmmlu_railway_and_automotive_engineering",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Railway-and-Automotive-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_real_estate": {
"task": "kmmlu_real_estate",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Real-Estate",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_refrigerating_machinery": {
"task": "kmmlu_refrigerating_machinery",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Refrigerating-Machinery",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_social_welfare": {
"task": "kmmlu_social_welfare",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Social-Welfare",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_taxation": {
"task": "kmmlu_taxation",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Taxation",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kmmlu_telecommunications_and_wireless_technology": {
"task": "kmmlu_telecommunications_and_wireless_technology",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Telecommunications-and-Wireless-Technology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kobest_boolq": {
"task": "kobest_boolq",
"dataset_path": "skt/kobest_v1",
"dataset_name": "boolq",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"doc_to_text": "{{paragraph}} 질문: {{question}} 답변: ",
"doc_to_target": "{{label}}",
"unsafe_code": false,
"doc_to_choice": [
"아니오",
"예"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": null,
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{paragraph}} 질문: {{question}} 답변: ",
"doc_to_choice": [
"아니오",
"예"
],
"doc_to_target": "{{label}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "f1",
"aggregation": "def macro_f1_score(items):\n from sklearn.metrics import f1_score\n\n unzipped_list = list(zip(*items))\n golds = unzipped_list[0]\n preds = unzipped_list[1]\n fscore = f1_score(golds, preds, average=\"macro\")\n return fscore\n",
"average": "macro",
"hf_evaluate": true,
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kobest_copa": {
"task": "kobest_copa",
"dataset_path": "skt/kobest_v1",
"dataset_name": "copa",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"doc_to_text": "def copa_doc_to_text(doc: dict) -> str:\n connector = {\"원인\": \" 왜냐하면\", \"결과\": \" 그래서\"}[doc[\"question\"].strip()]\n return f\"\"\"{doc[\"premise\"]} {connector}\"\"\"\n",
"doc_to_target": "def copa_doc_to_target(doc: dict) -> str:\n correct_choice = doc[\"alternative_1\"] if doc[\"label\"] == 0 else doc[\"alternative_2\"]\n return f\"\"\"{correct_choice}\"\"\"\n",
"unsafe_code": false,
"doc_to_choice": "def copa_doc_to_choice(doc: dict) -> list:\n return [f\"\"\"{doc[\"alternative_1\"]}\"\"\", f\"\"\"{doc[\"alternative_2\"]}\"\"\"]\n",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": null,
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "<function copa_doc_to_text at 0x77cb4011d8a0>",
"doc_to_choice": "<function copa_doc_to_choice at 0x77cb4011e3e0>",
"doc_to_target": "<function copa_doc_to_target at 0x77cb4011dee0>",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "f1",
"aggregation": "def macro_f1_score(items):\n from sklearn.metrics import f1_score\n\n unzipped_list = list(zip(*items))\n golds = unzipped_list[0]\n preds = unzipped_list[1]\n fscore = f1_score(golds, preds, average=\"macro\")\n return fscore\n",
"average": "macro",
"hf_evaluate": true,
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"kobest_hellaswag": {
"task": "kobest_hellaswag",
"dataset_path": "skt/kobest_v1",
"dataset_name": "hellaswag",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"process_docs": "def hellaswag_process_doc(doc: Dataset) -> Dataset:\n def preprocessor(dataset):\n return {\n \"query\": f\"\"\"문장: {dataset[\"context\"]}\"\"\",\n \"choices\": [\n dataset[\"ending_1\"],\n dataset[\"ending_2\"],\n dataset[\"ending_3\"],\n dataset[\"ending_4\"],\n ],\n \"gold\": int(dataset[\"label\"]),\n }\n\n return doc.map(preprocessor)\n",
"doc_to_text": "{{query}}",
"doc_to_target": "{{label}}",
"unsafe_code": false,
"doc_to_choice": "choices",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": null,
"process_docs": "<function hellaswag_process_doc at 0x77cb4011f060>",
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{query}}",
"doc_to_choice": "choices",
"doc_to_target": "{{label}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "acc_norm",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "f1",
"aggregation": "def macro_f1_score(items):\n from sklearn.metrics import f1_score\n\n unzipped_list = list(zip(*items))\n golds = unzipped_list[0]\n preds = unzipped_list[1]\n fscore = f1_score(golds, preds, average=\"macro\")\n return fscore\n",
"average": "macro",
"hf_evaluate": true,
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_abstract_algebra": {
"task": "mmlu_abstract_algebra",
"task_alias": "abstract_algebra",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "abstract_algebra",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about abstract algebra.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_anatomy": {
"task": "mmlu_anatomy",
"task_alias": "anatomy",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "anatomy",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about anatomy.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_astronomy": {
"task": "mmlu_astronomy",
"task_alias": "astronomy",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "astronomy",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about astronomy.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_business_ethics": {
"task": "mmlu_business_ethics",
"task_alias": "business_ethics",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "business_ethics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about business ethics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_clinical_knowledge": {
"task": "mmlu_clinical_knowledge",
"task_alias": "clinical_knowledge",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "clinical_knowledge",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about clinical knowledge.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_college_biology": {
"task": "mmlu_college_biology",
"task_alias": "college_biology",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "college_biology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about college biology.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_college_chemistry": {
"task": "mmlu_college_chemistry",
"task_alias": "college_chemistry",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "college_chemistry",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about college chemistry.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_college_computer_science": {
"task": "mmlu_college_computer_science",
"task_alias": "college_computer_science",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "college_computer_science",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about college computer science.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_college_mathematics": {
"task": "mmlu_college_mathematics",
"task_alias": "college_mathematics",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "college_mathematics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about college mathematics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_college_medicine": {
"task": "mmlu_college_medicine",
"task_alias": "college_medicine",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "college_medicine",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about college medicine.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_college_physics": {
"task": "mmlu_college_physics",
"task_alias": "college_physics",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "college_physics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about college physics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_computer_security": {
"task": "mmlu_computer_security",
"task_alias": "computer_security",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "computer_security",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about computer security.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_conceptual_physics": {
"task": "mmlu_conceptual_physics",
"task_alias": "conceptual_physics",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "conceptual_physics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about conceptual physics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_econometrics": {
"task": "mmlu_econometrics",
"task_alias": "econometrics",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "econometrics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about econometrics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_electrical_engineering": {
"task": "mmlu_electrical_engineering",
"task_alias": "electrical_engineering",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "electrical_engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about electrical engineering.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_elementary_mathematics": {
"task": "mmlu_elementary_mathematics",
"task_alias": "elementary_mathematics",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "elementary_mathematics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about elementary mathematics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_formal_logic": {
"task": "mmlu_formal_logic",
"task_alias": "formal_logic",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "formal_logic",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about formal logic.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_global_facts": {
"task": "mmlu_global_facts",
"task_alias": "global_facts",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "global_facts",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about global facts.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_high_school_biology": {
"task": "mmlu_high_school_biology",
"task_alias": "high_school_biology",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_biology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school biology.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_high_school_chemistry": {
"task": "mmlu_high_school_chemistry",
"task_alias": "high_school_chemistry",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_chemistry",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school chemistry.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_high_school_computer_science": {
"task": "mmlu_high_school_computer_science",
"task_alias": "high_school_computer_science",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_computer_science",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school computer science.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_high_school_european_history": {
"task": "mmlu_high_school_european_history",
"task_alias": "high_school_european_history",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_european_history",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school european history.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_high_school_geography": {
"task": "mmlu_high_school_geography",
"task_alias": "high_school_geography",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_geography",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school geography.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_high_school_government_and_politics": {
"task": "mmlu_high_school_government_and_politics",
"task_alias": "high_school_government_and_politics",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_government_and_politics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school government and politics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_high_school_macroeconomics": {
"task": "mmlu_high_school_macroeconomics",
"task_alias": "high_school_macroeconomics",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_macroeconomics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school macroeconomics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_high_school_mathematics": {
"task": "mmlu_high_school_mathematics",
"task_alias": "high_school_mathematics",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_mathematics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school mathematics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_high_school_microeconomics": {
"task": "mmlu_high_school_microeconomics",
"task_alias": "high_school_microeconomics",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_microeconomics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school microeconomics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_high_school_physics": {
"task": "mmlu_high_school_physics",
"task_alias": "high_school_physics",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_physics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school physics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_high_school_psychology": {
"task": "mmlu_high_school_psychology",
"task_alias": "high_school_psychology",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_psychology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school psychology.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_high_school_statistics": {
"task": "mmlu_high_school_statistics",
"task_alias": "high_school_statistics",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_statistics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school statistics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_high_school_us_history": {
"task": "mmlu_high_school_us_history",
"task_alias": "high_school_us_history",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_us_history",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school us history.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_high_school_world_history": {
"task": "mmlu_high_school_world_history",
"task_alias": "high_school_world_history",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_world_history",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school world history.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_human_aging": {
"task": "mmlu_human_aging",
"task_alias": "human_aging",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "human_aging",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about human aging.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_human_sexuality": {
"task": "mmlu_human_sexuality",
"task_alias": "human_sexuality",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "human_sexuality",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about human sexuality.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_international_law": {
"task": "mmlu_international_law",
"task_alias": "international_law",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "international_law",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about international law.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_jurisprudence": {
"task": "mmlu_jurisprudence",
"task_alias": "jurisprudence",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "jurisprudence",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about jurisprudence.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_logical_fallacies": {
"task": "mmlu_logical_fallacies",
"task_alias": "logical_fallacies",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "logical_fallacies",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about logical fallacies.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_machine_learning": {
"task": "mmlu_machine_learning",
"task_alias": "machine_learning",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "machine_learning",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about machine learning.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_management": {
"task": "mmlu_management",
"task_alias": "management",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "management",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about management.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_marketing": {
"task": "mmlu_marketing",
"task_alias": "marketing",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "marketing",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about marketing.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_medical_genetics": {
"task": "mmlu_medical_genetics",
"task_alias": "medical_genetics",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "medical_genetics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about medical genetics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_miscellaneous": {
"task": "mmlu_miscellaneous",
"task_alias": "miscellaneous",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "miscellaneous",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about miscellaneous.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_moral_disputes": {
"task": "mmlu_moral_disputes",
"task_alias": "moral_disputes",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "moral_disputes",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about moral disputes.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_moral_scenarios": {
"task": "mmlu_moral_scenarios",
"task_alias": "moral_scenarios",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "moral_scenarios",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about moral scenarios.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_nutrition": {
"task": "mmlu_nutrition",
"task_alias": "nutrition",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "nutrition",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about nutrition.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_philosophy": {
"task": "mmlu_philosophy",
"task_alias": "philosophy",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "philosophy",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about philosophy.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_prehistory": {
"task": "mmlu_prehistory",
"task_alias": "prehistory",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "prehistory",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about prehistory.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_professional_accounting": {
"task": "mmlu_professional_accounting",
"task_alias": "professional_accounting",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "professional_accounting",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about professional accounting.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_professional_law": {
"task": "mmlu_professional_law",
"task_alias": "professional_law",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "professional_law",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about professional law.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_professional_medicine": {
"task": "mmlu_professional_medicine",
"task_alias": "professional_medicine",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "professional_medicine",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about professional medicine.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_professional_psychology": {
"task": "mmlu_professional_psychology",
"task_alias": "professional_psychology",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "professional_psychology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about professional psychology.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_public_relations": {
"task": "mmlu_public_relations",
"task_alias": "public_relations",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "public_relations",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about public relations.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_security_studies": {
"task": "mmlu_security_studies",
"task_alias": "security_studies",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "security_studies",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about security studies.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_sociology": {
"task": "mmlu_sociology",
"task_alias": "sociology",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "sociology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about sociology.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_us_foreign_policy": {
"task": "mmlu_us_foreign_policy",
"task_alias": "us_foreign_policy",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "us_foreign_policy",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about us foreign policy.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_virology": {
"task": "mmlu_virology",
"task_alias": "virology",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "virology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about virology.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"mmlu_world_religions": {
"task": "mmlu_world_religions",
"task_alias": "world_religions",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "world_religions",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about world religions.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
},
"winogrande": {
"task": "winogrande",
"dataset_path": "allenai/winogrande",
"dataset_name": "winogrande_xl",
"training_split": "train",
"validation_split": "validation",
"doc_to_text": "def doc_to_text(doc):\n answer_to_num = {\"1\": 0, \"2\": 1}\n return answer_to_num[doc[\"answer\"]]\n",
"doc_to_target": "def doc_to_target(doc):\n idx = doc[\"sentence\"].index(\"_\") + 1\n return doc[\"sentence\"][idx:].strip()\n",
"unsafe_code": false,
"doc_to_choice": "def doc_to_choice(doc):\n idx = doc[\"sentence\"].index(\"_\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"sentence\"][:idx] + opt for opt in options]\n",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": null,
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "<function doc_to_text at 0x77cb4f17d620>",
"doc_to_choice": "<function doc_to_choice at 0x77cb4f17dc60>",
"doc_to_target": "<function doc_to_target at 0x77cb4f17d9e0>",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": true,
"doc_to_decontamination_query": "sentence",
"metadata": {
"version": 1.0,
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
}
}
},
"versions": {
"arc_challenge": 1.0,
"arc_easy": 1.0,
"hellaswag": 1.0,
"kmmlu": 2.0,
"kmmlu_accounting": 2.0,
"kmmlu_agricultural_sciences": 2.0,
"kmmlu_applied_science": 2.0,
"kmmlu_aviation_engineering_and_maintenance": 2.0,
"kmmlu_biology": 2.0,
"kmmlu_chemical_engineering": 2.0,
"kmmlu_chemistry": 2.0,
"kmmlu_civil_engineering": 2.0,
"kmmlu_computer_science": 2.0,
"kmmlu_construction": 2.0,
"kmmlu_criminal_law": 2.0,
"kmmlu_ecology": 2.0,
"kmmlu_economics": 2.0,
"kmmlu_education": 2.0,
"kmmlu_electrical_engineering": 2.0,
"kmmlu_electronics_engineering": 2.0,
"kmmlu_energy_management": 2.0,
"kmmlu_environmental_science": 2.0,
"kmmlu_fashion": 2.0,
"kmmlu_food_processing": 2.0,
"kmmlu_gas_technology_and_engineering": 2.0,
"kmmlu_geomatics": 2.0,
"kmmlu_health": 2.0,
"kmmlu_humss": 2.0,
"kmmlu_industrial_engineer": 2.0,
"kmmlu_information_technology": 2.0,
"kmmlu_interior_architecture_and_design": 2.0,
"kmmlu_korean_history": 2.0,
"kmmlu_law": 2.0,
"kmmlu_machine_design_and_manufacturing": 2.0,
"kmmlu_management": 2.0,
"kmmlu_maritime_engineering": 2.0,
"kmmlu_marketing": 2.0,
"kmmlu_materials_engineering": 2.0,
"kmmlu_math": 2.0,
"kmmlu_mechanical_engineering": 2.0,
"kmmlu_nondestructive_testing": 2.0,
"kmmlu_other": 2.0,
"kmmlu_patent": 2.0,
"kmmlu_political_science_and_sociology": 2.0,
"kmmlu_psychology": 2.0,
"kmmlu_public_safety": 2.0,
"kmmlu_railway_and_automotive_engineering": 2.0,
"kmmlu_real_estate": 2.0,
"kmmlu_refrigerating_machinery": 2.0,
"kmmlu_social_welfare": 2.0,
"kmmlu_stem": 2.0,
"kmmlu_taxation": 2.0,
"kmmlu_telecommunications_and_wireless_technology": 2.0,
"kobest_boolq": 1.0,
"kobest_copa": 1.0,
"kobest_hellaswag": 1.0,
"mmlu": 2,
"mmlu_abstract_algebra": 1.0,
"mmlu_anatomy": 1.0,
"mmlu_astronomy": 1.0,
"mmlu_business_ethics": 1.0,
"mmlu_clinical_knowledge": 1.0,
"mmlu_college_biology": 1.0,
"mmlu_college_chemistry": 1.0,
"mmlu_college_computer_science": 1.0,
"mmlu_college_mathematics": 1.0,
"mmlu_college_medicine": 1.0,
"mmlu_college_physics": 1.0,
"mmlu_computer_security": 1.0,
"mmlu_conceptual_physics": 1.0,
"mmlu_econometrics": 1.0,
"mmlu_electrical_engineering": 1.0,
"mmlu_elementary_mathematics": 1.0,
"mmlu_formal_logic": 1.0,
"mmlu_global_facts": 1.0,
"mmlu_high_school_biology": 1.0,
"mmlu_high_school_chemistry": 1.0,
"mmlu_high_school_computer_science": 1.0,
"mmlu_high_school_european_history": 1.0,
"mmlu_high_school_geography": 1.0,
"mmlu_high_school_government_and_politics": 1.0,
"mmlu_high_school_macroeconomics": 1.0,
"mmlu_high_school_mathematics": 1.0,
"mmlu_high_school_microeconomics": 1.0,
"mmlu_high_school_physics": 1.0,
"mmlu_high_school_psychology": 1.0,
"mmlu_high_school_statistics": 1.0,
"mmlu_high_school_us_history": 1.0,
"mmlu_high_school_world_history": 1.0,
"mmlu_human_aging": 1.0,
"mmlu_human_sexuality": 1.0,
"mmlu_humanities": 2,
"mmlu_international_law": 1.0,
"mmlu_jurisprudence": 1.0,
"mmlu_logical_fallacies": 1.0,
"mmlu_machine_learning": 1.0,
"mmlu_management": 1.0,
"mmlu_marketing": 1.0,
"mmlu_medical_genetics": 1.0,
"mmlu_miscellaneous": 1.0,
"mmlu_moral_disputes": 1.0,
"mmlu_moral_scenarios": 1.0,
"mmlu_nutrition": 1.0,
"mmlu_other": 2,
"mmlu_philosophy": 1.0,
"mmlu_prehistory": 1.0,
"mmlu_professional_accounting": 1.0,
"mmlu_professional_law": 1.0,
"mmlu_professional_medicine": 1.0,
"mmlu_professional_psychology": 1.0,
"mmlu_public_relations": 1.0,
"mmlu_security_studies": 1.0,
"mmlu_social_sciences": 2,
"mmlu_sociology": 1.0,
"mmlu_stem": 2,
"mmlu_us_foreign_policy": 1.0,
"mmlu_virology": 1.0,
"mmlu_world_religions": 1.0,
"winogrande": 1.0
},
"n-shot": {
"arc_challenge": 0,
"arc_easy": 0,
"hellaswag": 0,
"kmmlu_accounting": 0,
"kmmlu_agricultural_sciences": 0,
"kmmlu_aviation_engineering_and_maintenance": 0,
"kmmlu_biology": 0,
"kmmlu_chemical_engineering": 0,
"kmmlu_chemistry": 0,
"kmmlu_civil_engineering": 0,
"kmmlu_computer_science": 0,
"kmmlu_construction": 0,
"kmmlu_criminal_law": 0,
"kmmlu_ecology": 0,
"kmmlu_economics": 0,
"kmmlu_education": 0,
"kmmlu_electrical_engineering": 0,
"kmmlu_electronics_engineering": 0,
"kmmlu_energy_management": 0,
"kmmlu_environmental_science": 0,
"kmmlu_fashion": 0,
"kmmlu_food_processing": 0,
"kmmlu_gas_technology_and_engineering": 0,
"kmmlu_geomatics": 0,
"kmmlu_health": 0,
"kmmlu_industrial_engineer": 0,
"kmmlu_information_technology": 0,
"kmmlu_interior_architecture_and_design": 0,
"kmmlu_korean_history": 0,
"kmmlu_law": 0,
"kmmlu_machine_design_and_manufacturing": 0,
"kmmlu_management": 0,
"kmmlu_maritime_engineering": 0,
"kmmlu_marketing": 0,
"kmmlu_materials_engineering": 0,
"kmmlu_math": 0,
"kmmlu_mechanical_engineering": 0,
"kmmlu_nondestructive_testing": 0,
"kmmlu_patent": 0,
"kmmlu_political_science_and_sociology": 0,
"kmmlu_psychology": 0,
"kmmlu_public_safety": 0,
"kmmlu_railway_and_automotive_engineering": 0,
"kmmlu_real_estate": 0,
"kmmlu_refrigerating_machinery": 0,
"kmmlu_social_welfare": 0,
"kmmlu_taxation": 0,
"kmmlu_telecommunications_and_wireless_technology": 0,
"kobest_boolq": 0,
"kobest_copa": 0,
"kobest_hellaswag": 0,
"mmlu_abstract_algebra": 0,
"mmlu_anatomy": 0,
"mmlu_astronomy": 0,
"mmlu_business_ethics": 0,
"mmlu_clinical_knowledge": 0,
"mmlu_college_biology": 0,
"mmlu_college_chemistry": 0,
"mmlu_college_computer_science": 0,
"mmlu_college_mathematics": 0,
"mmlu_college_medicine": 0,
"mmlu_college_physics": 0,
"mmlu_computer_security": 0,
"mmlu_conceptual_physics": 0,
"mmlu_econometrics": 0,
"mmlu_electrical_engineering": 0,
"mmlu_elementary_mathematics": 0,
"mmlu_formal_logic": 0,
"mmlu_global_facts": 0,
"mmlu_high_school_biology": 0,
"mmlu_high_school_chemistry": 0,
"mmlu_high_school_computer_science": 0,
"mmlu_high_school_european_history": 0,
"mmlu_high_school_geography": 0,
"mmlu_high_school_government_and_politics": 0,
"mmlu_high_school_macroeconomics": 0,
"mmlu_high_school_mathematics": 0,
"mmlu_high_school_microeconomics": 0,
"mmlu_high_school_physics": 0,
"mmlu_high_school_psychology": 0,
"mmlu_high_school_statistics": 0,
"mmlu_high_school_us_history": 0,
"mmlu_high_school_world_history": 0,
"mmlu_human_aging": 0,
"mmlu_human_sexuality": 0,
"mmlu_international_law": 0,
"mmlu_jurisprudence": 0,
"mmlu_logical_fallacies": 0,
"mmlu_machine_learning": 0,
"mmlu_management": 0,
"mmlu_marketing": 0,
"mmlu_medical_genetics": 0,
"mmlu_miscellaneous": 0,
"mmlu_moral_disputes": 0,
"mmlu_moral_scenarios": 0,
"mmlu_nutrition": 0,
"mmlu_philosophy": 0,
"mmlu_prehistory": 0,
"mmlu_professional_accounting": 0,
"mmlu_professional_law": 0,
"mmlu_professional_medicine": 0,
"mmlu_professional_psychology": 0,
"mmlu_public_relations": 0,
"mmlu_security_studies": 0,
"mmlu_sociology": 0,
"mmlu_us_foreign_policy": 0,
"mmlu_virology": 0,
"mmlu_world_religions": 0,
"winogrande": 0
},
"higher_is_better": {
"arc_challenge": {
"acc": true,
"acc_norm": true
},
"arc_easy": {
"acc": true,
"acc_norm": true
},
"hellaswag": {
"acc": true,
"acc_norm": true
},
"kmmlu": {
"acc": true
},
"kmmlu_accounting": {
"acc": true
},
"kmmlu_agricultural_sciences": {
"acc": true
},
"kmmlu_applied_science": {
"acc": true
},
"kmmlu_aviation_engineering_and_maintenance": {
"acc": true
},
"kmmlu_biology": {
"acc": true
},
"kmmlu_chemical_engineering": {
"acc": true
},
"kmmlu_chemistry": {
"acc": true
},
"kmmlu_civil_engineering": {
"acc": true
},
"kmmlu_computer_science": {
"acc": true
},
"kmmlu_construction": {
"acc": true
},
"kmmlu_criminal_law": {
"acc": true
},
"kmmlu_ecology": {
"acc": true
},
"kmmlu_economics": {
"acc": true
},
"kmmlu_education": {
"acc": true
},
"kmmlu_electrical_engineering": {
"acc": true
},
"kmmlu_electronics_engineering": {
"acc": true
},
"kmmlu_energy_management": {
"acc": true
},
"kmmlu_environmental_science": {
"acc": true
},
"kmmlu_fashion": {
"acc": true
},
"kmmlu_food_processing": {
"acc": true
},
"kmmlu_gas_technology_and_engineering": {
"acc": true
},
"kmmlu_geomatics": {
"acc": true
},
"kmmlu_health": {
"acc": true
},
"kmmlu_humss": {
"acc": true
},
"kmmlu_industrial_engineer": {
"acc": true
},
"kmmlu_information_technology": {
"acc": true
},
"kmmlu_interior_architecture_and_design": {
"acc": true
},
"kmmlu_korean_history": {
"acc": true
},
"kmmlu_law": {
"acc": true
},
"kmmlu_machine_design_and_manufacturing": {
"acc": true
},
"kmmlu_management": {
"acc": true
},
"kmmlu_maritime_engineering": {
"acc": true
},
"kmmlu_marketing": {
"acc": true
},
"kmmlu_materials_engineering": {
"acc": true
},
"kmmlu_math": {
"acc": true
},
"kmmlu_mechanical_engineering": {
"acc": true
},
"kmmlu_nondestructive_testing": {
"acc": true
},
"kmmlu_other": {
"acc": true
},
"kmmlu_patent": {
"acc": true
},
"kmmlu_political_science_and_sociology": {
"acc": true
},
"kmmlu_psychology": {
"acc": true
},
"kmmlu_public_safety": {
"acc": true
},
"kmmlu_railway_and_automotive_engineering": {
"acc": true
},
"kmmlu_real_estate": {
"acc": true
},
"kmmlu_refrigerating_machinery": {
"acc": true
},
"kmmlu_social_welfare": {
"acc": true
},
"kmmlu_stem": {
"acc": true
},
"kmmlu_taxation": {
"acc": true
},
"kmmlu_telecommunications_and_wireless_technology": {
"acc": true
},
"kobest_boolq": {
"acc": true,
"f1": true
},
"kobest_copa": {
"acc": true,
"f1": true
},
"kobest_hellaswag": {
"acc": true,
"acc_norm": true,
"f1": true
},
"mmlu": {
"acc": true
},
"mmlu_abstract_algebra": {
"acc": true
},
"mmlu_anatomy": {
"acc": true
},
"mmlu_astronomy": {
"acc": true
},
"mmlu_business_ethics": {
"acc": true
},
"mmlu_clinical_knowledge": {
"acc": true
},
"mmlu_college_biology": {
"acc": true
},
"mmlu_college_chemistry": {
"acc": true
},
"mmlu_college_computer_science": {
"acc": true
},
"mmlu_college_mathematics": {
"acc": true
},
"mmlu_college_medicine": {
"acc": true
},
"mmlu_college_physics": {
"acc": true
},
"mmlu_computer_security": {
"acc": true
},
"mmlu_conceptual_physics": {
"acc": true
},
"mmlu_econometrics": {
"acc": true
},
"mmlu_electrical_engineering": {
"acc": true
},
"mmlu_elementary_mathematics": {
"acc": true
},
"mmlu_formal_logic": {
"acc": true
},
"mmlu_global_facts": {
"acc": true
},
"mmlu_high_school_biology": {
"acc": true
},
"mmlu_high_school_chemistry": {
"acc": true
},
"mmlu_high_school_computer_science": {
"acc": true
},
"mmlu_high_school_european_history": {
"acc": true
},
"mmlu_high_school_geography": {
"acc": true
},
"mmlu_high_school_government_and_politics": {
"acc": true
},
"mmlu_high_school_macroeconomics": {
"acc": true
},
"mmlu_high_school_mathematics": {
"acc": true
},
"mmlu_high_school_microeconomics": {
"acc": true
},
"mmlu_high_school_physics": {
"acc": true
},
"mmlu_high_school_psychology": {
"acc": true
},
"mmlu_high_school_statistics": {
"acc": true
},
"mmlu_high_school_us_history": {
"acc": true
},
"mmlu_high_school_world_history": {
"acc": true
},
"mmlu_human_aging": {
"acc": true
},
"mmlu_human_sexuality": {
"acc": true
},
"mmlu_humanities": {
"acc": true
},
"mmlu_international_law": {
"acc": true
},
"mmlu_jurisprudence": {
"acc": true
},
"mmlu_logical_fallacies": {
"acc": true
},
"mmlu_machine_learning": {
"acc": true
},
"mmlu_management": {
"acc": true
},
"mmlu_marketing": {
"acc": true
},
"mmlu_medical_genetics": {
"acc": true
},
"mmlu_miscellaneous": {
"acc": true
},
"mmlu_moral_disputes": {
"acc": true
},
"mmlu_moral_scenarios": {
"acc": true
},
"mmlu_nutrition": {
"acc": true
},
"mmlu_other": {
"acc": true
},
"mmlu_philosophy": {
"acc": true
},
"mmlu_prehistory": {
"acc": true
},
"mmlu_professional_accounting": {
"acc": true
},
"mmlu_professional_law": {
"acc": true
},
"mmlu_professional_medicine": {
"acc": true
},
"mmlu_professional_psychology": {
"acc": true
},
"mmlu_public_relations": {
"acc": true
},
"mmlu_security_studies": {
"acc": true
},
"mmlu_social_sciences": {
"acc": true
},
"mmlu_sociology": {
"acc": true
},
"mmlu_stem": {
"acc": true
},
"mmlu_us_foreign_policy": {
"acc": true
},
"mmlu_virology": {
"acc": true
},
"mmlu_world_religions": {
"acc": true
},
"winogrande": {
"acc": true
}
},
"n-samples": {
"kobest_hellaswag": {
"original": 500,
"effective": 400
},
"kobest_copa": {
"original": 1000,
"effective": 400
},
"kobest_boolq": {
"original": 1404,
"effective": 400
},
"kmmlu_biology": {
"original": 1000,
"effective": 400
},
"kmmlu_chemical_engineering": {
"original": 1000,
"effective": 400
},
"kmmlu_chemistry": {
"original": 600,
"effective": 400
},
"kmmlu_civil_engineering": {
"original": 1000,
"effective": 400
},
"kmmlu_computer_science": {
"original": 1000,
"effective": 400
},
"kmmlu_ecology": {
"original": 1000,
"effective": 400
},
"kmmlu_electrical_engineering": {
"original": 1000,
"effective": 400
},
"kmmlu_information_technology": {
"original": 1000,
"effective": 400
},
"kmmlu_materials_engineering": {
"original": 1000,
"effective": 400
},
"kmmlu_math": {
"original": 300,
"effective": 300
},
"kmmlu_mechanical_engineering": {
"original": 1000,
"effective": 400
},
"kmmlu_agricultural_sciences": {
"original": 1000,
"effective": 400
},
"kmmlu_construction": {
"original": 1000,
"effective": 400
},
"kmmlu_fashion": {
"original": 1000,
"effective": 400
},
"kmmlu_food_processing": {
"original": 1000,
"effective": 400
},
"kmmlu_health": {
"original": 100,
"effective": 100
},
"kmmlu_interior_architecture_and_design": {
"original": 1000,
"effective": 400
},
"kmmlu_marketing": {
"original": 1000,
"effective": 400
},
"kmmlu_patent": {
"original": 100,
"effective": 100
},
"kmmlu_public_safety": {
"original": 1000,
"effective": 400
},
"kmmlu_real_estate": {
"original": 200,
"effective": 200
},
"kmmlu_refrigerating_machinery": {
"original": 1000,
"effective": 400
},
"kmmlu_aviation_engineering_and_maintenance": {
"original": 1000,
"effective": 400
},
"kmmlu_electronics_engineering": {
"original": 1000,
"effective": 400
},
"kmmlu_energy_management": {
"original": 1000,
"effective": 400
},
"kmmlu_environmental_science": {
"original": 1000,
"effective": 400
},
"kmmlu_gas_technology_and_engineering": {
"original": 1000,
"effective": 400
},
"kmmlu_geomatics": {
"original": 1000,
"effective": 400
},
"kmmlu_industrial_engineer": {
"original": 1000,
"effective": 400
},
"kmmlu_machine_design_and_manufacturing": {
"original": 1000,
"effective": 400
},
"kmmlu_maritime_engineering": {
"original": 600,
"effective": 400
},
"kmmlu_nondestructive_testing": {
"original": 1000,
"effective": 400
},
"kmmlu_railway_and_automotive_engineering": {
"original": 1000,
"effective": 400
},
"kmmlu_telecommunications_and_wireless_technology": {
"original": 1000,
"effective": 400
},
"kmmlu_accounting": {
"original": 100,
"effective": 100
},
"kmmlu_criminal_law": {
"original": 200,
"effective": 200
},
"kmmlu_economics": {
"original": 130,
"effective": 130
},
"kmmlu_education": {
"original": 100,
"effective": 100
},
"kmmlu_korean_history": {
"original": 100,
"effective": 100
},
"kmmlu_law": {
"original": 1000,
"effective": 400
},
"kmmlu_management": {
"original": 1000,
"effective": 400
},
"kmmlu_political_science_and_sociology": {
"original": 300,
"effective": 300
},
"kmmlu_psychology": {
"original": 1000,
"effective": 400
},
"kmmlu_social_welfare": {
"original": 1000,
"effective": 400
},
"kmmlu_taxation": {
"original": 200,
"effective": 200
},
"winogrande": {
"original": 1267,
"effective": 400
},
"arc_challenge": {
"original": 1172,
"effective": 400
},
"arc_easy": {
"original": 2376,
"effective": 400
},
"hellaswag": {
"original": 10042,
"effective": 400
},
"mmlu_abstract_algebra": {
"original": 100,
"effective": 100
},
"mmlu_anatomy": {
"original": 135,
"effective": 135
},
"mmlu_astronomy": {
"original": 152,
"effective": 152
},
"mmlu_college_biology": {
"original": 144,
"effective": 144
},
"mmlu_college_chemistry": {
"original": 100,
"effective": 100
},
"mmlu_college_computer_science": {
"original": 100,
"effective": 100
},
"mmlu_college_mathematics": {
"original": 100,
"effective": 100
},
"mmlu_college_physics": {
"original": 102,
"effective": 102
},
"mmlu_computer_security": {
"original": 100,
"effective": 100
},
"mmlu_conceptual_physics": {
"original": 235,
"effective": 235
},
"mmlu_electrical_engineering": {
"original": 145,
"effective": 145
},
"mmlu_elementary_mathematics": {
"original": 378,
"effective": 378
},
"mmlu_high_school_biology": {
"original": 310,
"effective": 310
},
"mmlu_high_school_chemistry": {
"original": 203,
"effective": 203
},
"mmlu_high_school_computer_science": {
"original": 100,
"effective": 100
},
"mmlu_high_school_mathematics": {
"original": 270,
"effective": 270
},
"mmlu_high_school_physics": {
"original": 151,
"effective": 151
},
"mmlu_high_school_statistics": {
"original": 216,
"effective": 216
},
"mmlu_machine_learning": {
"original": 112,
"effective": 112
},
"mmlu_business_ethics": {
"original": 100,
"effective": 100
},
"mmlu_clinical_knowledge": {
"original": 265,
"effective": 265
},
"mmlu_college_medicine": {
"original": 173,
"effective": 173
},
"mmlu_global_facts": {
"original": 100,
"effective": 100
},
"mmlu_human_aging": {
"original": 223,
"effective": 223
},
"mmlu_management": {
"original": 103,
"effective": 103
},
"mmlu_marketing": {
"original": 234,
"effective": 234
},
"mmlu_medical_genetics": {
"original": 100,
"effective": 100
},
"mmlu_miscellaneous": {
"original": 783,
"effective": 400
},
"mmlu_nutrition": {
"original": 306,
"effective": 306
},
"mmlu_professional_accounting": {
"original": 282,
"effective": 282
},
"mmlu_professional_medicine": {
"original": 272,
"effective": 272
},
"mmlu_virology": {
"original": 166,
"effective": 166
},
"mmlu_econometrics": {
"original": 114,
"effective": 114
},
"mmlu_high_school_geography": {
"original": 198,
"effective": 198
},
"mmlu_high_school_government_and_politics": {
"original": 193,
"effective": 193
},
"mmlu_high_school_macroeconomics": {
"original": 390,
"effective": 390
},
"mmlu_high_school_microeconomics": {
"original": 238,
"effective": 238
},
"mmlu_high_school_psychology": {
"original": 545,
"effective": 400
},
"mmlu_human_sexuality": {
"original": 131,
"effective": 131
},
"mmlu_professional_psychology": {
"original": 612,
"effective": 400
},
"mmlu_public_relations": {
"original": 110,
"effective": 110
},
"mmlu_security_studies": {
"original": 245,
"effective": 245
},
"mmlu_sociology": {
"original": 201,
"effective": 201
},
"mmlu_us_foreign_policy": {
"original": 100,
"effective": 100
},
"mmlu_formal_logic": {
"original": 126,
"effective": 126
},
"mmlu_high_school_european_history": {
"original": 165,
"effective": 165
},
"mmlu_high_school_us_history": {
"original": 204,
"effective": 204
},
"mmlu_high_school_world_history": {
"original": 237,
"effective": 237
},
"mmlu_international_law": {
"original": 121,
"effective": 121
},
"mmlu_jurisprudence": {
"original": 108,
"effective": 108
},
"mmlu_logical_fallacies": {
"original": 163,
"effective": 163
},
"mmlu_moral_disputes": {
"original": 346,
"effective": 346
},
"mmlu_moral_scenarios": {
"original": 895,
"effective": 400
},
"mmlu_philosophy": {
"original": 311,
"effective": 311
},
"mmlu_prehistory": {
"original": 324,
"effective": 324
},
"mmlu_professional_law": {
"original": 1534,
"effective": 400
},
"mmlu_world_religions": {
"original": 171,
"effective": 171
}
},
"config": {
"model": "hf",
"model_args": {
"pretrained": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"trust_remote_code": true
},
"model_num_parameters": 4022468096,
"model_dtype": "torch.bfloat16",
"model_revision": "main",
"model_sha": "",
"batch_size": "12",
"batch_sizes": [],
"device": "cuda:0",
"use_cache": null,
"limit": 400.0,
"bootstrap_iters": 100000,
"gen_kwargs": {},
"random_seed": 0,
"numpy_seed": 1234,
"torch_seed": 1234,
"fewshot_seed": 1234
},
"git_hash": "0ce43af",
"date": 1775962096.959724,
"pretty_env_info": "PyTorch version: 2.9.0+cu128\nIs debug build: False\nCUDA used to build PyTorch: 12.8\nROCM used to build PyTorch: N/A\n\nOS: Ubuntu 22.04.5 LTS (x86_64)\nGCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0\nClang version: Could not collect\nCMake version: version 4.1.0\nLibc version: glibc-2.35\n\nPython version: 3.11.14 | packaged by conda-forge | (main, Oct 13 2025, 14:09:32) [GCC 14.3.0] (64-bit runtime)\nPython platform: Linux-6.8.0-90-generic-x86_64-with-glibc2.35\nIs CUDA available: True\nCUDA runtime version: 12.8.93\nCUDA_MODULE_LOADING set to: \nGPU models and configuration: GPU 0: NVIDIA RTX PRO 6000 Blackwell Workstation Edition\nNvidia driver version: 590.48.01\ncuDNN version: Probably one of the following:\n/usr/lib/x86_64-linux-gnu/libcudnn.so.9.8.0\n/usr/lib/x86_64-linux-gnu/libcudnn_adv.so.9.8.0\n/usr/lib/x86_64-linux-gnu/libcudnn_cnn.so.9.8.0\n/usr/lib/x86_64-linux-gnu/libcudnn_engines_precompiled.so.9.8.0\n/usr/lib/x86_64-linux-gnu/libcudnn_engines_runtime_compiled.so.9.8.0\n/usr/lib/x86_64-linux-gnu/libcudnn_graph.so.9.8.0\n/usr/lib/x86_64-linux-gnu/libcudnn_heuristic.so.9.8.0\n/usr/lib/x86_64-linux-gnu/libcudnn_ops.so.9.8.0\nIs XPU available: False\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nAddress sizes: 43 bits physical, 48 bits virtual\nByte Order: Little Endian\nCPU(s): 192\nOn-line CPU(s) list: 0-191\nVendor ID: AuthenticAMD\nModel name: AMD EPYC 7642 48-Core Processor\nCPU family: 23\nModel: 49\nThread(s) per core: 2\nCore(s) per socket: 48\nSocket(s): 2\nStepping: 0\nFrequency boost: enabled\nCPU max MHz: 2300.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 4600.15\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid aperfmperf rapl pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_llc mwaitx cpb cat_l3 cdp_l3 hw_pstate ssbd mba ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr rdpru wbnoinvd amd_ppin arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif v_spec_ctrl umip rdpid overflow_recov succor smca sev sev_es ibpb_exit_to_user\nVirtualization: AMD-V\nL1d cache: 3 MiB (96 instances)\nL1i cache: 3 MiB (96 instances)\nL2 cache: 48 MiB (96 instances)\nL3 cache: 512 MiB (32 instances)\nNUMA node(s): 2\nNUMA node0 CPU(s): 0-47,96-143\nNUMA node1 CPU(s): 48-95,144-191\nVulnerability Gather data sampling: Not affected\nVulnerability Itlb multihit: Not affected\nVulnerability L1tf: Not affected\nVulnerability Mds: Not affected\nVulnerability Meltdown: Not affected\nVulnerability Mmio stale data: Not affected\nVulnerability Reg file data sampling: Not affected\nVulnerability Retbleed: Mitigation; untrained return thunk; SMT enabled with STIBP protection\nVulnerability Spec rs
"transformers_version": "5.5.3",
"lm_eval_version": "0.4.11",
"upper_git_hash": null,
"tokenizer_pad_token": [
"<|PAD_TOKEN|>",
"151669"
],
"tokenizer_eos_token": [
"<|endoftext|>",
"151643"
],
"tokenizer_bos_token": [
null,
"None"
],
"eot_token_id": 151643,
"max_length": 32768,
"task_hashes": {},
"model_source": "hf",
"model_name": "/home/unsloth/scp_stage1_cpt/artifacts/cpt_full_96gb_qwen3_4b/checkpoints",
"model_name_sanitized": "__home__unsloth__scp_stage1_cpt__artifacts__cpt_full_96gb_qwen3_4b__checkpoints",
"system_instruction": null,
"system_instruction_sha": null,
"fewshot_as_multiturn": null,
"chat_template": null,
"chat_template_sha": null,
"total_evaluation_time_seconds": "580.1511918641627"
},
"base": {
"results": {
"arc_challenge": {
"alias": "arc_challenge",
"acc,none": 0.4525,
"acc_stderr,none": 0.024918098926991643,
"acc_norm,none": 0.4975,
"acc_norm_stderr,none": 0.0250309958227734
},
"arc_easy": {
"alias": "arc_easy",
"acc,none": 0.7625,
"acc_stderr,none": 0.02130420258115865,
"acc_norm,none": 0.755,
"acc_norm_stderr,none": 0.02153129097913246
},
"hellaswag": {
"alias": "hellaswag",
"acc,none": 0.4925,
"acc_stderr,none": 0.025028492535438325,
"acc_norm,none": 0.6225,
"acc_norm_stderr,none": 0.024268431488608636
},
"kmmlu": {
"acc,none": 0.47556707712248864,
"acc_stderr,none": 0.003918983222456166,
"alias": "kmmlu"
},
"kmmlu_applied_science": {
"acc,none": 0.45875,
"acc_stderr,none": 0.007101063857525891,
"alias": " - kmmlu_applied_science"
},
"kmmlu_aviation_engineering_and_maintenance": {
"alias": " - kmmlu_aviation_engineering_and_maintenance",
"acc,none": 0.4475,
"acc_stderr,none": 0.024892941194307603
},
"kmmlu_electronics_engineering": {
"alias": " - kmmlu_electronics_engineering",
"acc,none": 0.65,
"acc_stderr,none": 0.023878346647046
},
"kmmlu_energy_management": {
"alias": " - kmmlu_energy_management",
"acc,none": 0.4,
"acc_stderr,none": 0.02452557357939856
},
"kmmlu_environmental_science": {
"alias": " - kmmlu_environmental_science",
"acc,none": 0.3875,
"acc_stderr,none": 0.02438947500927543
},
"kmmlu_gas_technology_and_engineering": {
"alias": " - kmmlu_gas_technology_and_engineering",
"acc,none": 0.3775,
"acc_stderr,none": 0.02426843148860864
},
"kmmlu_geomatics": {
"alias": " - kmmlu_geomatics",
"acc,none": 0.4325,
"acc_stderr,none": 0.024802162065186362
},
"kmmlu_industrial_engineer": {
"alias": " - kmmlu_industrial_engineer",
"acc,none": 0.4275,
"acc_stderr,none": 0.024766769210836766
},
"kmmlu_machine_design_and_manufacturing": {
"alias": " - kmmlu_machine_design_and_manufacturing",
"acc,none": 0.52,
"acc_stderr,none": 0.025011275652681887
},
"kmmlu_maritime_engineering": {
"alias": " - kmmlu_maritime_engineering",
"acc,none": 0.405,
"acc_stderr,none": 0.024575340657273674
},
"kmmlu_nondestructive_testing": {
"alias": " - kmmlu_nondestructive_testing",
"acc,none": 0.4825,
"acc_stderr,none": 0.025015972341295333
},
"kmmlu_railway_and_automotive_engineering": {
"alias": " - kmmlu_railway_and_automotive_engineering",
"acc,none": 0.3875,
"acc_stderr,none": 0.02438947500927542
},
"kmmlu_telecommunications_and_wireless_technology": {
"alias": " - kmmlu_telecommunications_and_wireless_technology",
"acc,none": 0.5875,
"acc_stderr,none": 0.024645036407943802
},
"kmmlu_humss": {
"acc,none": 0.4805860805860806,
"acc_stderr,none": 0.009419825503999339,
"alias": " - kmmlu_humss"
},
"kmmlu_accounting": {
"alias": " - kmmlu_accounting",
"acc,none": 0.49,
"acc_stderr,none": 0.05024183937956912
},
"kmmlu_criminal_law": {
"alias": " - kmmlu_criminal_law",
"acc,none": 0.39,
"acc_stderr,none": 0.03457567623250011
},
"kmmlu_economics": {
"alias": " - kmmlu_economics",
"acc,none": 0.5615384615384615,
"acc_stderr,none": 0.04368784779071991
},
"kmmlu_education": {
"alias": " - kmmlu_education",
"acc,none": 0.65,
"acc_stderr,none": 0.047937248544110196
},
"kmmlu_korean_history": {
"alias": " - kmmlu_korean_history",
"acc,none": 0.24,
"acc_stderr,none": 0.04292346959909284
},
"kmmlu_law": {
"alias": " - kmmlu_law",
"acc,none": 0.3875,
"acc_stderr,none": 0.024389475009275435
},
"kmmlu_management": {
"alias": " - kmmlu_management",
"acc,none": 0.53,
"acc_stderr,none": 0.02498621173652297
},
"kmmlu_political_science_and_sociology": {
"alias": " - kmmlu_political_science_and_sociology",
"acc,none": 0.5466666666666666,
"acc_stderr,none": 0.028789526978043094
},
"kmmlu_psychology": {
"alias": " - kmmlu_psychology",
"acc,none": 0.4275,
"acc_stderr,none": 0.02476676921083677
},
"kmmlu_social_welfare": {
"alias": " - kmmlu_social_welfare",
"acc,none": 0.585,
"acc_stderr,none": 0.02466695454685353
},
"kmmlu_taxation": {
"alias": " - kmmlu_taxation",
"acc,none": 0.435,
"acc_stderr,none": 0.03514328173714407
},
"kmmlu_other": {
"acc,none": 0.4772222222222222,
"acc_stderr,none": 0.008073884461069719,
"alias": " - kmmlu_other"
},
"kmmlu_agricultural_sciences": {
"alias": " - kmmlu_agricultural_sciences",
"acc,none": 0.3625,
"acc_stderr,none": 0.024066207238097725
},
"kmmlu_construction": {
"alias": " - kmmlu_construction",
"acc,none": 0.3925,
"acc_stderr,none": 0.024445927747963316
},
"kmmlu_fashion": {
"alias": " - kmmlu_fashion",
"acc,none": 0.4575,
"acc_stderr,none": 0.024940719189394073
},
"kmmlu_food_processing": {
"alias": " - kmmlu_food_processing",
"acc,none": 0.39,
"acc_stderr,none": 0.024418038445046374
},
"kmmlu_health": {
"alias": " - kmmlu_health",
"acc,none": 0.63,
"acc_stderr,none": 0.048523658709391
},
"kmmlu_interior_architecture_and_design": {
"alias": " - kmmlu_interior_architecture_and_design",
"acc,none": 0.6025,
"acc_stderr,none": 0.024499693108404712
},
"kmmlu_marketing": {
"alias": " - kmmlu_marketing",
"acc,none": 0.76,
"acc_stderr,none": 0.021380899352993952
},
"kmmlu_patent": {
"alias": " - kmmlu_patent",
"acc,none": 0.46,
"acc_stderr,none": 0.05009082659620332
},
"kmmlu_public_safety": {
"alias": " - kmmlu_public_safety",
"acc,none": 0.4025,
"acc_stderr,none": 0.024550788746396206
},
"kmmlu_real_estate": {
"alias": " - kmmlu_real_estate",
"acc,none": 0.485,
"acc_stderr,none": 0.03542810683297719
},
"kmmlu_refrigerating_machinery": {
"alias": " - kmmlu_refrigerating_machinery",
"acc,none": 0.4125,
"acc_stderr,none": 0.024645036407943802
},
"kmmlu_stem": {
"acc,none": 0.4897674418604651,
"acc_stderr,none": 0.007312394370135803,
"alias": " - kmmlu_stem"
},
"kmmlu_biology": {
"alias": " - kmmlu_biology",
"acc,none": 0.3225,
"acc_stderr,none": 0.023400926978618723
},
"kmmlu_chemical_engineering": {
"alias": " - kmmlu_chemical_engineering",
"acc,none": 0.4875,
"acc_stderr,none": 0.025023485209500245
},
"kmmlu_chemistry": {
"alias": " - kmmlu_chemistry",
"acc,none": 0.5175,
"acc_stderr,none": 0.02501597234129533
},
"kmmlu_civil_engineering": {
"alias": " - kmmlu_civil_engineering",
"acc,none": 0.3825,
"acc_stderr,none": 0.024330316186072946
},
"kmmlu_computer_science": {
"alias": " - kmmlu_computer_science",
"acc,none": 0.75,
"acc_stderr,none": 0.021677749238103
},
"kmmlu_ecology": {
"alias": " - kmmlu_ecology",
"acc,none": 0.5425,
"acc_stderr,none": 0.024940719189394077
},
"kmmlu_electrical_engineering": {
"alias": " - kmmlu_electrical_engineering",
"acc,none": 0.355,
"acc_stderr,none": 0.023955629410456463
},
"kmmlu_information_technology": {
"alias": " - kmmlu_information_technology",
"acc,none": 0.75,
"acc_stderr,none": 0.021677749238103
},
"kmmlu_materials_engineering": {
"alias": " - kmmlu_materials_engineering",
"acc,none": 0.495,
"acc_stderr,none": 0.025030057119361453
},
"kmmlu_math": {
"alias": " - kmmlu_math",
"acc,none": 0.3333333333333333,
"acc_stderr,none": 0.027262027336984396
},
"kmmlu_mechanical_engineering": {
"alias": " - kmmlu_mechanical_engineering",
"acc,none": 0.4125,
"acc_stderr,none": 0.024645036407943802
},
"kobest_boolq": {
"alias": "kobest_boolq",
"acc,none": 0.6675,
"acc_stderr,none": 0.023584952830141535,
"f1,none": 0.6247575383530242,
"f1_stderr,none": "N/A"
},
"kobest_copa": {
"alias": "kobest_copa",
"acc,none": 0.6475,
"acc_stderr,none": 0.023917346710791564,
"f1,none": 0.6473920138042275,
"f1_stderr,none": "N/A"
},
"kobest_hellaswag": {
"alias": "kobest_hellaswag",
"acc,none": 0.44,
"acc_stderr,none": 0.02485042976789583,
"f1,none": 0.4328647077786627,
"f1_stderr,none": "N/A",
"acc_norm,none": 0.5825,
"acc_norm_stderr,none": 0.024688218756390913
},
"mmlu": {
"acc,none": 0.7404266255461321,
"acc_stderr,none": 0.003869340083262106,
"alias": "mmlu"
},
"mmlu_humanities": {
"acc,none": 0.6931079323797139,
"acc_stderr,none": 0.0077779673157217745,
"alias": " - humanities"
},
"mmlu_formal_logic": {
"alias": " - formal_logic",
"acc,none": 0.5793650793650794,
"acc_stderr,none": 0.04415438226743745
},
"mmlu_high_school_european_history": {
"alias": " - high_school_european_history",
"acc,none": 0.7818181818181819,
"acc_stderr,none": 0.03225078108306289
},
"mmlu_high_school_us_history": {
"alias": " - high_school_us_history",
"acc,none": 0.8284313725490197,
"acc_stderr,none": 0.02646056956124065
},
"mmlu_high_school_world_history": {
"alias": " - high_school_world_history",
"acc,none": 0.8438818565400844,
"acc_stderr,none": 0.023627159460318684
},
"mmlu_international_law": {
"alias": " - international_law",
"acc,none": 0.8016528925619835,
"acc_stderr,none": 0.03640118271990946
},
"mmlu_jurisprudence": {
"alias": " - jurisprudence",
"acc,none": 0.7962962962962963,
"acc_stderr,none": 0.03893542518824847
},
"mmlu_logical_fallacies": {
"alias": " - logical_fallacies",
"acc,none": 0.8404907975460123,
"acc_stderr,none": 0.02876748172598387
},
"mmlu_moral_disputes": {
"alias": " - moral_disputes",
"acc,none": 0.7543352601156069,
"acc_stderr,none": 0.023176298203992
},
"mmlu_moral_scenarios": {
"alias": " - moral_scenarios",
"acc,none": 0.3475,
"acc_stderr,none": 0.023838625698390636
},
"mmlu_philosophy": {
"alias": " - philosophy",
"acc,none": 0.7588424437299035,
"acc_stderr,none": 0.02429659403476343
},
"mmlu_prehistory": {
"alias": " - prehistory",
"acc,none": 0.7870370370370371,
"acc_stderr,none": 0.02277971908873339
},
"mmlu_professional_law": {
"alias": " - professional_law",
"acc,none": 0.53,
"acc_stderr,none": 0.02498621173652297
},
"mmlu_world_religions": {
"alias": " - world_religions",
"acc,none": 0.8070175438596491,
"acc_stderr,none": 0.030267457554898458
},
"mmlu_other": {
"acc,none": 0.7437591776798825,
"acc_stderr,none": 0.008056333552095894,
"alias": " - other"
},
"mmlu_business_ethics": {
"alias": " - business_ethics",
"acc,none": 0.75,
"acc_stderr,none": 0.04351941398892446
},
"mmlu_clinical_knowledge": {
"alias": " - clinical_knowledge",
"acc,none": 0.7773584905660378,
"acc_stderr,none": 0.0256042334708991
},
"mmlu_college_medicine": {
"alias": " - college_medicine",
"acc,none": 0.7341040462427746,
"acc_stderr,none": 0.03368762932259431
},
"mmlu_global_facts": {
"alias": " - global_facts",
"acc,none": 0.43,
"acc_stderr,none": 0.04975698519562429
},
"mmlu_human_aging": {
"alias": " - human_aging",
"acc,none": 0.7488789237668162,
"acc_stderr,none": 0.02910522083322461
},
"mmlu_management": {
"alias": " - management",
"acc,none": 0.8932038834951457,
"acc_stderr,none": 0.030581088928331356
},
"mmlu_marketing": {
"alias": " - marketing",
"acc,none": 0.9145299145299145,
"acc_stderr,none": 0.018315891685625862
},
"mmlu_medical_genetics": {
"alias": " - medical_genetics",
"acc,none": 0.8,
"acc_stderr,none": 0.04020151261036846
},
"mmlu_miscellaneous": {
"alias": " - miscellaneous",
"acc,none": 0.82,
"acc_stderr,none": 0.01923342954415769
},
"mmlu_nutrition": {
"alias": " - nutrition",
"acc,none": 0.7745098039215687,
"acc_stderr,none": 0.023929155517351277
},
"mmlu_professional_accounting": {
"alias": " - professional_accounting",
"acc,none": 0.5709219858156028,
"acc_stderr,none": 0.02952591430255856
},
"mmlu_professional_medicine": {
"alias": " - professional_medicine",
"acc,none": 0.7757352941176471,
"acc_stderr,none": 0.025336848563332365
},
"mmlu_virology": {
"alias": " - virology",
"acc,none": 0.5120481927710844,
"acc_stderr,none": 0.03891364495835817
},
"mmlu_social_sciences": {
"acc,none": 0.8202205882352941,
"acc_stderr,none": 0.007248431086566561,
"alias": " - social sciences"
},
"mmlu_econometrics": {
"alias": " - econometrics",
"acc,none": 0.6578947368421053,
"acc_stderr,none": 0.04462917535336937
},
"mmlu_high_school_geography": {
"alias": " - high_school_geography",
"acc,none": 0.8737373737373737,
"acc_stderr,none": 0.02366435940288024
},
"mmlu_high_school_government_and_politics": {
"alias": " - high_school_government_and_politics",
"acc,none": 0.8756476683937824,
"acc_stderr,none": 0.023814477086593556
},
"mmlu_high_school_macroeconomics": {
"alias": " - high_school_macroeconomics",
"acc,none": 0.8076923076923077,
"acc_stderr,none": 0.019982347208637292
},
"mmlu_high_school_microeconomics": {
"alias": " - high_school_microeconomics",
"acc,none": 0.8991596638655462,
"acc_stderr,none": 0.019559663430480802
},
"mmlu_high_school_psychology": {
"alias": " - high_school_psychology",
"acc,none": 0.9025,
"acc_stderr,none": 0.0148504449187799
},
"mmlu_human_sexuality": {
"alias": " - human_sexuality",
"acc,none": 0.7862595419847328,
"acc_stderr,none": 0.035954616117746904
},
"mmlu_professional_psychology": {
"alias": " - professional_psychology",
"acc,none": 0.7475,
"acc_stderr,none": 0.0217495282695941
},
"mmlu_public_relations": {
"alias": " - public_relations",
"acc,none": 0.6818181818181818,
"acc_stderr,none": 0.04461272175910509
},
"mmlu_security_studies": {
"alias": " - security_studies",
"acc,none": 0.7673469387755102,
"acc_stderr,none": 0.02704925791589618
},
"mmlu_sociology": {
"alias": " - sociology",
"acc,none": 0.845771144278607,
"acc_stderr,none": 0.02553843336857833
},
"mmlu_us_foreign_policy": {
"alias": " - us_foreign_policy",
"acc,none": 0.89,
"acc_stderr,none": 0.03144660377352203
},
"mmlu_stem": {
"acc,none": 0.7148747224865207,
"acc_stderr,none": 0.007751851248299227,
"alias": " - stem"
},
"mmlu_abstract_algebra": {
"alias": " - abstract_algebra",
"acc,none": 0.47,
"acc_stderr,none": 0.050161355804659205
},
"mmlu_anatomy": {
"alias": " - anatomy",
"acc,none": 0.6888888888888889,
"acc_stderr,none": 0.03999262876617723
},
"mmlu_astronomy": {
"alias": " - astronomy",
"acc,none": 0.8421052631578947,
"acc_stderr,none": 0.02967416752010141
},
"mmlu_college_biology": {
"alias": " - college_biology",
"acc,none": 0.8402777777777778,
"acc_stderr,none": 0.030635578972093267
},
"mmlu_college_chemistry": {
"alias": " - college_chemistry",
"acc,none": 0.56,
"acc_stderr,none": 0.049888765156985884
},
"mmlu_college_computer_science": {
"alias": " - college_computer_science",
"acc,none": 0.66,
"acc_stderr,none": 0.04760952285695237
},
"mmlu_college_mathematics": {
"alias": " - college_mathematics",
"acc,none": 0.52,
"acc_stderr,none": 0.050211673156867795
},
"mmlu_college_physics": {
"alias": " - college_physics",
"acc,none": 0.5686274509803921,
"acc_stderr,none": 0.04928099597287534
},
"mmlu_computer_security": {
"alias": " - computer_security",
"acc,none": 0.83,
"acc_stderr,none": 0.0377525168068637
},
"mmlu_conceptual_physics": {
"alias": " - conceptual_physics",
"acc,none": 0.7957446808510639,
"acc_stderr,none": 0.026355158413349428
},
"mmlu_electrical_engineering": {
"alias": " - electrical_engineering",
"acc,none": 0.7517241379310344,
"acc_stderr,none": 0.036001056927277716
},
"mmlu_elementary_mathematics": {
"alias": " - elementary_mathematics",
"acc,none": 0.7116402116402116,
"acc_stderr,none": 0.023330654054535903
},
"mmlu_high_school_biology": {
"alias": " - high_school_biology",
"acc,none": 0.9161290322580645,
"acc_stderr,none": 0.015769027496775653
},
"mmlu_high_school_chemistry": {
"alias": " - high_school_chemistry",
"acc,none": 0.7192118226600985,
"acc_stderr,none": 0.03161856335358611
},
"mmlu_high_school_computer_science": {
"alias": " - high_school_computer_science",
"acc,none": 0.87,
"acc_stderr,none": 0.03379976689896309
},
"mmlu_high_school_mathematics": {
"alias": " - high_school_mathematics",
"acc,none": 0.5222222222222223,
"acc_stderr,none": 0.030455413985678408
},
"mmlu_high_school_physics": {
"alias": " - high_school_physics",
"acc,none": 0.6754966887417219,
"acc_stderr,none": 0.038227469376587525
},
"mmlu_high_school_statistics": {
"alias": " - high_school_statistics",
"acc,none": 0.7222222222222222,
"acc_stderr,none": 0.030546745264953185
},
"mmlu_machine_learning": {
"alias": " - machine_learning",
"acc,none": 0.6160714285714286,
"acc_stderr,none": 0.04616143075028546
},
"winogrande": {
"alias": "winogrande",
"acc,none": 0.7375,
"acc_stderr,none": 0.022027196108925243
}
},
"groups": {
"kmmlu": {
"acc,none": 0.47556707712248864,
"acc_stderr,none": 0.003918983222456166,
"alias": "kmmlu"
},
"kmmlu_applied_science": {
"acc,none": 0.45875,
"acc_stderr,none": 0.007101063857525891,
"alias": " - kmmlu_applied_science"
},
"kmmlu_humss": {
"acc,none": 0.4805860805860806,
"acc_stderr,none": 0.009419825503999339,
"alias": " - kmmlu_humss"
},
"kmmlu_other": {
"acc,none": 0.4772222222222222,
"acc_stderr,none": 0.008073884461069719,
"alias": " - kmmlu_other"
},
"kmmlu_stem": {
"acc,none": 0.4897674418604651,
"acc_stderr,none": 0.007312394370135803,
"alias": " - kmmlu_stem"
},
"mmlu": {
"acc,none": 0.7404266255461321,
"acc_stderr,none": 0.003869340083262106,
"alias": "mmlu"
},
"mmlu_humanities": {
"acc,none": 0.6931079323797139,
"acc_stderr,none": 0.0077779673157217745,
"alias": " - humanities"
},
"mmlu_other": {
"acc,none": 0.7437591776798825,
"acc_stderr,none": 0.008056333552095894,
"alias": " - other"
},
"mmlu_social_sciences": {
"acc,none": 0.8202205882352941,
"acc_stderr,none": 0.007248431086566561,
"alias": " - social sciences"
},
"mmlu_stem": {
"acc,none": 0.7148747224865207,
"acc_stderr,none": 0.007751851248299227,
"alias": " - stem"
}
},
"group_subtasks": {
"mmlu_humanities": [
"mmlu_formal_logic",
"mmlu_high_school_european_history",
"mmlu_high_school_us_history",
"mmlu_high_school_world_history",
"mmlu_international_law",
"mmlu_jurisprudence",
"mmlu_logical_fallacies",
"mmlu_moral_disputes",
"mmlu_moral_scenarios",
"mmlu_philosophy",
"mmlu_prehistory",
"mmlu_professional_law",
"mmlu_world_religions"
],
"mmlu_social_sciences": [
"mmlu_econometrics",
"mmlu_high_school_geography",
"mmlu_high_school_government_and_politics",
"mmlu_high_school_macroeconomics",
"mmlu_high_school_microeconomics",
"mmlu_high_school_psychology",
"mmlu_human_sexuality",
"mmlu_professional_psychology",
"mmlu_public_relations",
"mmlu_security_studies",
"mmlu_sociology",
"mmlu_us_foreign_policy"
],
"mmlu_other": [
"mmlu_business_ethics",
"mmlu_clinical_knowledge",
"mmlu_college_medicine",
"mmlu_global_facts",
"mmlu_human_aging",
"mmlu_management",
"mmlu_marketing",
"mmlu_medical_genetics",
"mmlu_miscellaneous",
"mmlu_nutrition",
"mmlu_professional_accounting",
"mmlu_professional_medicine",
"mmlu_virology"
],
"mmlu_stem": [
"mmlu_abstract_algebra",
"mmlu_anatomy",
"mmlu_astronomy",
"mmlu_college_biology",
"mmlu_college_chemistry",
"mmlu_college_computer_science",
"mmlu_college_mathematics",
"mmlu_college_physics",
"mmlu_computer_security",
"mmlu_conceptual_physics",
"mmlu_electrical_engineering",
"mmlu_elementary_mathematics",
"mmlu_high_school_biology",
"mmlu_high_school_chemistry",
"mmlu_high_school_computer_science",
"mmlu_high_school_mathematics",
"mmlu_high_school_physics",
"mmlu_high_school_statistics",
"mmlu_machine_learning"
],
"mmlu": [
"mmlu_stem",
"mmlu_other",
"mmlu_social_sciences",
"mmlu_humanities"
],
"hellaswag": [],
"arc_easy": [],
"arc_challenge": [],
"winogrande": [],
"kmmlu_humss": [
"kmmlu_accounting",
"kmmlu_criminal_law",
"kmmlu_economics",
"kmmlu_education",
"kmmlu_korean_history",
"kmmlu_law",
"kmmlu_management",
"kmmlu_political_science_and_sociology",
"kmmlu_psychology",
"kmmlu_social_welfare",
"kmmlu_taxation"
],
"kmmlu_applied_science": [
"kmmlu_aviation_engineering_and_maintenance",
"kmmlu_electronics_engineering",
"kmmlu_energy_management",
"kmmlu_environmental_science",
"kmmlu_gas_technology_and_engineering",
"kmmlu_geomatics",
"kmmlu_industrial_engineer",
"kmmlu_machine_design_and_manufacturing",
"kmmlu_maritime_engineering",
"kmmlu_nondestructive_testing",
"kmmlu_railway_and_automotive_engineering",
"kmmlu_telecommunications_and_wireless_technology"
],
"kmmlu_other": [
"kmmlu_agricultural_sciences",
"kmmlu_construction",
"kmmlu_fashion",
"kmmlu_food_processing",
"kmmlu_health",
"kmmlu_interior_architecture_and_design",
"kmmlu_marketing",
"kmmlu_patent",
"kmmlu_public_safety",
"kmmlu_real_estate",
"kmmlu_refrigerating_machinery"
],
"kmmlu_stem": [
"kmmlu_biology",
"kmmlu_chemical_engineering",
"kmmlu_chemistry",
"kmmlu_civil_engineering",
"kmmlu_computer_science",
"kmmlu_ecology",
"kmmlu_electrical_engineering",
"kmmlu_information_technology",
"kmmlu_materials_engineering",
"kmmlu_math",
"kmmlu_mechanical_engineering"
],
"kmmlu": [
"kmmlu_stem",
"kmmlu_other",
"kmmlu_applied_science",
"kmmlu_humss"
],
"kobest_boolq": [],
"kobest_copa": [],
"kobest_hellaswag": []
},
"configs": {
"arc_challenge": {
"task": "arc_challenge",
"tag": [
"ai2_arc"
],
"dataset_path": "allenai/ai2_arc",
"dataset_name": "ARC-Challenge",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": "{{choices.label.index(answerKey)}}",
"unsafe_code": false,
"doc_to_choice": "{{choices.text}}",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": null,
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_choice": "{{choices.text}}",
"doc_to_target": "{{choices.label.index(answerKey)}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "acc_norm",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": true,
"doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"arc_easy": {
"task": "arc_easy",
"tag": [
"ai2_arc"
],
"dataset_path": "allenai/ai2_arc",
"dataset_name": "ARC-Easy",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": "{{choices.label.index(answerKey)}}",
"unsafe_code": false,
"doc_to_choice": "{{choices.text}}",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": null,
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_choice": "{{choices.text}}",
"doc_to_target": "{{choices.label.index(answerKey)}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "acc_norm",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": true,
"doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"hellaswag": {
"task": "hellaswag",
"tag": [
"multiple_choice"
],
"dataset_path": "Rowan/hellaswag",
"training_split": "train",
"validation_split": "validation",
"process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc):\n ctx = doc[\"ctx_a\"] + \" \" + doc[\"ctx_b\"].capitalize()\n out_doc = {\n \"query\": preprocess(doc[\"activity_label\"] + \": \" + ctx),\n \"choices\": [preprocess(ending) for ending in doc[\"endings\"]],\n \"gold\": int(doc[\"label\"]),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
"doc_to_text": "{{query}}",
"doc_to_target": "{{label}}",
"unsafe_code": false,
"doc_to_choice": "choices",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": null,
"process_docs": "<function process_docs at 0x73f0acb5a700>",
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{query}}",
"doc_to_choice": "choices",
"doc_to_target": "{{label}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "acc_norm",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_accounting": {
"task": "kmmlu_accounting",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Accounting",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_agricultural_sciences": {
"task": "kmmlu_agricultural_sciences",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Agricultural-Sciences",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_aviation_engineering_and_maintenance": {
"task": "kmmlu_aviation_engineering_and_maintenance",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Aviation-Engineering-and-Maintenance",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_biology": {
"task": "kmmlu_biology",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Biology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_chemical_engineering": {
"task": "kmmlu_chemical_engineering",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Chemical-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_chemistry": {
"task": "kmmlu_chemistry",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Chemistry",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_civil_engineering": {
"task": "kmmlu_civil_engineering",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Civil-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_computer_science": {
"task": "kmmlu_computer_science",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Computer-Science",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_construction": {
"task": "kmmlu_construction",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Construction",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_criminal_law": {
"task": "kmmlu_criminal_law",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Criminal-Law",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_ecology": {
"task": "kmmlu_ecology",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Ecology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_economics": {
"task": "kmmlu_economics",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Economics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_education": {
"task": "kmmlu_education",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Education",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_electrical_engineering": {
"task": "kmmlu_electrical_engineering",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Electrical-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_electronics_engineering": {
"task": "kmmlu_electronics_engineering",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Electronics-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_energy_management": {
"task": "kmmlu_energy_management",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Energy-Management",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_environmental_science": {
"task": "kmmlu_environmental_science",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Environmental-Science",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_fashion": {
"task": "kmmlu_fashion",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Fashion",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_food_processing": {
"task": "kmmlu_food_processing",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Food-Processing",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_gas_technology_and_engineering": {
"task": "kmmlu_gas_technology_and_engineering",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Gas-Technology-and-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_geomatics": {
"task": "kmmlu_geomatics",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Geomatics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_health": {
"task": "kmmlu_health",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Health",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_industrial_engineer": {
"task": "kmmlu_industrial_engineer",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Industrial-Engineer",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_information_technology": {
"task": "kmmlu_information_technology",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Information-Technology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_interior_architecture_and_design": {
"task": "kmmlu_interior_architecture_and_design",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Interior-Architecture-and-Design",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_korean_history": {
"task": "kmmlu_korean_history",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Korean-History",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_law": {
"task": "kmmlu_law",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Law",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_machine_design_and_manufacturing": {
"task": "kmmlu_machine_design_and_manufacturing",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Machine-Design-and-Manufacturing",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_management": {
"task": "kmmlu_management",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Management",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_maritime_engineering": {
"task": "kmmlu_maritime_engineering",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Maritime-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_marketing": {
"task": "kmmlu_marketing",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Marketing",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_materials_engineering": {
"task": "kmmlu_materials_engineering",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Materials-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_math": {
"task": "kmmlu_math",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Math",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_mechanical_engineering": {
"task": "kmmlu_mechanical_engineering",
"tag": "kmmlu_stem_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Mechanical-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_nondestructive_testing": {
"task": "kmmlu_nondestructive_testing",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Nondestructive-Testing",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_patent": {
"task": "kmmlu_patent",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Patent",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_political_science_and_sociology": {
"task": "kmmlu_political_science_and_sociology",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Political-Science-and-Sociology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_psychology": {
"task": "kmmlu_psychology",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Psychology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_public_safety": {
"task": "kmmlu_public_safety",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Public-Safety",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_railway_and_automotive_engineering": {
"task": "kmmlu_railway_and_automotive_engineering",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Railway-and-Automotive-Engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_real_estate": {
"task": "kmmlu_real_estate",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Real-Estate",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_refrigerating_machinery": {
"task": "kmmlu_refrigerating_machinery",
"tag": "kmmlu_other_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Refrigerating-Machinery",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_social_welfare": {
"task": "kmmlu_social_welfare",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Social-Welfare",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_taxation": {
"task": "kmmlu_taxation",
"tag": "kmmlu_humss_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Taxation",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kmmlu_telecommunications_and_wireless_technology": {
"task": "kmmlu_telecommunications_and_wireless_technology",
"tag": "kmmlu_applied_science_tasks",
"dataset_path": "HAERAE-HUB/KMMLU",
"dataset_name": "Telecommunications-and-Wireless-Technology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_target": "{{answer-1}}",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n정답",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "{{answer-1}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 2.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kobest_boolq": {
"task": "kobest_boolq",
"dataset_path": "skt/kobest_v1",
"dataset_name": "boolq",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"doc_to_text": "{{paragraph}} 질문: {{question}} 답변: ",
"doc_to_target": "{{label}}",
"unsafe_code": false,
"doc_to_choice": [
"아니오",
"예"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": null,
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{paragraph}} 질문: {{question}} 답변: ",
"doc_to_choice": [
"아니오",
"예"
],
"doc_to_target": "{{label}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "f1",
"aggregation": "def macro_f1_score(items):\n from sklearn.metrics import f1_score\n\n unzipped_list = list(zip(*items))\n golds = unzipped_list[0]\n preds = unzipped_list[1]\n fscore = f1_score(golds, preds, average=\"macro\")\n return fscore\n",
"average": "macro",
"hf_evaluate": true,
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kobest_copa": {
"task": "kobest_copa",
"dataset_path": "skt/kobest_v1",
"dataset_name": "copa",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"doc_to_text": "def copa_doc_to_text(doc: dict) -> str:\n connector = {\"원인\": \" 왜냐하면\", \"결과\": \" 그래서\"}[doc[\"question\"].strip()]\n return f\"\"\"{doc[\"premise\"]} {connector}\"\"\"\n",
"doc_to_target": "def copa_doc_to_target(doc: dict) -> str:\n correct_choice = doc[\"alternative_1\"] if doc[\"label\"] == 0 else doc[\"alternative_2\"]\n return f\"\"\"{correct_choice}\"\"\"\n",
"unsafe_code": false,
"doc_to_choice": "def copa_doc_to_choice(doc: dict) -> list:\n return [f\"\"\"{doc[\"alternative_1\"]}\"\"\", f\"\"\"{doc[\"alternative_2\"]}\"\"\"]\n",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": null,
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "<function copa_doc_to_text at 0x73f083a685e0>",
"doc_to_choice": "<function copa_doc_to_choice at 0x73f083a69120>",
"doc_to_target": "<function copa_doc_to_target at 0x73f083a68c20>",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "f1",
"aggregation": "def macro_f1_score(items):\n from sklearn.metrics import f1_score\n\n unzipped_list = list(zip(*items))\n golds = unzipped_list[0]\n preds = unzipped_list[1]\n fscore = f1_score(golds, preds, average=\"macro\")\n return fscore\n",
"average": "macro",
"hf_evaluate": true,
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"kobest_hellaswag": {
"task": "kobest_hellaswag",
"dataset_path": "skt/kobest_v1",
"dataset_name": "hellaswag",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"process_docs": "def hellaswag_process_doc(doc: Dataset) -> Dataset:\n def preprocessor(dataset):\n return {\n \"query\": f\"\"\"문장: {dataset[\"context\"]}\"\"\",\n \"choices\": [\n dataset[\"ending_1\"],\n dataset[\"ending_2\"],\n dataset[\"ending_3\"],\n dataset[\"ending_4\"],\n ],\n \"gold\": int(dataset[\"label\"]),\n }\n\n return doc.map(preprocessor)\n",
"doc_to_text": "{{query}}",
"doc_to_target": "{{label}}",
"unsafe_code": false,
"doc_to_choice": "choices",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": null,
"process_docs": "<function hellaswag_process_doc at 0x73f083a69940>",
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{query}}",
"doc_to_choice": "choices",
"doc_to_target": "{{label}}",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "acc_norm",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "f1",
"aggregation": "def macro_f1_score(items):\n from sklearn.metrics import f1_score\n\n unzipped_list = list(zip(*items))\n golds = unzipped_list[0]\n preds = unzipped_list[1]\n fscore = f1_score(golds, preds, average=\"macro\")\n return fscore\n",
"average": "macro",
"hf_evaluate": true,
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_abstract_algebra": {
"task": "mmlu_abstract_algebra",
"task_alias": "abstract_algebra",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "abstract_algebra",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about abstract algebra.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_anatomy": {
"task": "mmlu_anatomy",
"task_alias": "anatomy",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "anatomy",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about anatomy.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_astronomy": {
"task": "mmlu_astronomy",
"task_alias": "astronomy",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "astronomy",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about astronomy.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_business_ethics": {
"task": "mmlu_business_ethics",
"task_alias": "business_ethics",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "business_ethics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about business ethics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_clinical_knowledge": {
"task": "mmlu_clinical_knowledge",
"task_alias": "clinical_knowledge",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "clinical_knowledge",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about clinical knowledge.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_college_biology": {
"task": "mmlu_college_biology",
"task_alias": "college_biology",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "college_biology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about college biology.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_college_chemistry": {
"task": "mmlu_college_chemistry",
"task_alias": "college_chemistry",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "college_chemistry",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about college chemistry.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_college_computer_science": {
"task": "mmlu_college_computer_science",
"task_alias": "college_computer_science",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "college_computer_science",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about college computer science.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_college_mathematics": {
"task": "mmlu_college_mathematics",
"task_alias": "college_mathematics",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "college_mathematics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about college mathematics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_college_medicine": {
"task": "mmlu_college_medicine",
"task_alias": "college_medicine",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "college_medicine",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about college medicine.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_college_physics": {
"task": "mmlu_college_physics",
"task_alias": "college_physics",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "college_physics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about college physics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_computer_security": {
"task": "mmlu_computer_security",
"task_alias": "computer_security",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "computer_security",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about computer security.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_conceptual_physics": {
"task": "mmlu_conceptual_physics",
"task_alias": "conceptual_physics",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "conceptual_physics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about conceptual physics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_econometrics": {
"task": "mmlu_econometrics",
"task_alias": "econometrics",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "econometrics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about econometrics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_electrical_engineering": {
"task": "mmlu_electrical_engineering",
"task_alias": "electrical_engineering",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "electrical_engineering",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about electrical engineering.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_elementary_mathematics": {
"task": "mmlu_elementary_mathematics",
"task_alias": "elementary_mathematics",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "elementary_mathematics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about elementary mathematics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_formal_logic": {
"task": "mmlu_formal_logic",
"task_alias": "formal_logic",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "formal_logic",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about formal logic.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_global_facts": {
"task": "mmlu_global_facts",
"task_alias": "global_facts",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "global_facts",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about global facts.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_high_school_biology": {
"task": "mmlu_high_school_biology",
"task_alias": "high_school_biology",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_biology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school biology.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_high_school_chemistry": {
"task": "mmlu_high_school_chemistry",
"task_alias": "high_school_chemistry",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_chemistry",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school chemistry.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_high_school_computer_science": {
"task": "mmlu_high_school_computer_science",
"task_alias": "high_school_computer_science",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_computer_science",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school computer science.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_high_school_european_history": {
"task": "mmlu_high_school_european_history",
"task_alias": "high_school_european_history",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_european_history",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school european history.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_high_school_geography": {
"task": "mmlu_high_school_geography",
"task_alias": "high_school_geography",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_geography",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school geography.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_high_school_government_and_politics": {
"task": "mmlu_high_school_government_and_politics",
"task_alias": "high_school_government_and_politics",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_government_and_politics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school government and politics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_high_school_macroeconomics": {
"task": "mmlu_high_school_macroeconomics",
"task_alias": "high_school_macroeconomics",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_macroeconomics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school macroeconomics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_high_school_mathematics": {
"task": "mmlu_high_school_mathematics",
"task_alias": "high_school_mathematics",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_mathematics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school mathematics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_high_school_microeconomics": {
"task": "mmlu_high_school_microeconomics",
"task_alias": "high_school_microeconomics",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_microeconomics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school microeconomics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_high_school_physics": {
"task": "mmlu_high_school_physics",
"task_alias": "high_school_physics",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_physics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school physics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_high_school_psychology": {
"task": "mmlu_high_school_psychology",
"task_alias": "high_school_psychology",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_psychology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school psychology.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_high_school_statistics": {
"task": "mmlu_high_school_statistics",
"task_alias": "high_school_statistics",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_statistics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school statistics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_high_school_us_history": {
"task": "mmlu_high_school_us_history",
"task_alias": "high_school_us_history",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_us_history",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school us history.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_high_school_world_history": {
"task": "mmlu_high_school_world_history",
"task_alias": "high_school_world_history",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "high_school_world_history",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about high school world history.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_human_aging": {
"task": "mmlu_human_aging",
"task_alias": "human_aging",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "human_aging",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about human aging.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_human_sexuality": {
"task": "mmlu_human_sexuality",
"task_alias": "human_sexuality",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "human_sexuality",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about human sexuality.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_international_law": {
"task": "mmlu_international_law",
"task_alias": "international_law",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "international_law",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about international law.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_jurisprudence": {
"task": "mmlu_jurisprudence",
"task_alias": "jurisprudence",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "jurisprudence",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about jurisprudence.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_logical_fallacies": {
"task": "mmlu_logical_fallacies",
"task_alias": "logical_fallacies",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "logical_fallacies",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about logical fallacies.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_machine_learning": {
"task": "mmlu_machine_learning",
"task_alias": "machine_learning",
"tag": "mmlu_stem_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "machine_learning",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about machine learning.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_management": {
"task": "mmlu_management",
"task_alias": "management",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "management",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about management.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_marketing": {
"task": "mmlu_marketing",
"task_alias": "marketing",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "marketing",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about marketing.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_medical_genetics": {
"task": "mmlu_medical_genetics",
"task_alias": "medical_genetics",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "medical_genetics",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about medical genetics.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_miscellaneous": {
"task": "mmlu_miscellaneous",
"task_alias": "miscellaneous",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "miscellaneous",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about miscellaneous.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_moral_disputes": {
"task": "mmlu_moral_disputes",
"task_alias": "moral_disputes",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "moral_disputes",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about moral disputes.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_moral_scenarios": {
"task": "mmlu_moral_scenarios",
"task_alias": "moral_scenarios",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "moral_scenarios",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about moral scenarios.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_nutrition": {
"task": "mmlu_nutrition",
"task_alias": "nutrition",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "nutrition",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about nutrition.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_philosophy": {
"task": "mmlu_philosophy",
"task_alias": "philosophy",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "philosophy",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about philosophy.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_prehistory": {
"task": "mmlu_prehistory",
"task_alias": "prehistory",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "prehistory",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about prehistory.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_professional_accounting": {
"task": "mmlu_professional_accounting",
"task_alias": "professional_accounting",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "professional_accounting",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about professional accounting.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_professional_law": {
"task": "mmlu_professional_law",
"task_alias": "professional_law",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "professional_law",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about professional law.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_professional_medicine": {
"task": "mmlu_professional_medicine",
"task_alias": "professional_medicine",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "professional_medicine",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about professional medicine.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_professional_psychology": {
"task": "mmlu_professional_psychology",
"task_alias": "professional_psychology",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "professional_psychology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about professional psychology.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_public_relations": {
"task": "mmlu_public_relations",
"task_alias": "public_relations",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "public_relations",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about public relations.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_security_studies": {
"task": "mmlu_security_studies",
"task_alias": "security_studies",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "security_studies",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about security studies.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_sociology": {
"task": "mmlu_sociology",
"task_alias": "sociology",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "sociology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about sociology.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_us_foreign_policy": {
"task": "mmlu_us_foreign_policy",
"task_alias": "us_foreign_policy",
"tag": "mmlu_social_sciences_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "us_foreign_policy",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about us foreign policy.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_virology": {
"task": "mmlu_virology",
"task_alias": "virology",
"tag": "mmlu_other_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "virology",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about virology.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"mmlu_world_religions": {
"task": "mmlu_world_religions",
"task_alias": "world_religions",
"tag": "mmlu_humanities_tasks",
"dataset_path": "cais/mmlu",
"dataset_name": "world_religions",
"test_split": "test",
"fewshot_split": "dev",
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_target": "answer",
"unsafe_code": false,
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "The following are multiple choice questions (with answers) about world religions.\n\n",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n",
"split": "dev",
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"doc_to_target": "answer",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
},
"winogrande": {
"task": "winogrande",
"dataset_path": "allenai/winogrande",
"dataset_name": "winogrande_xl",
"training_split": "train",
"validation_split": "validation",
"doc_to_text": "def doc_to_text(doc):\n answer_to_num = {\"1\": 0, \"2\": 1}\n return answer_to_num[doc[\"answer\"]]\n",
"doc_to_target": "def doc_to_target(doc):\n idx = doc[\"sentence\"].index(\"_\") + 1\n return doc[\"sentence\"][idx:].strip()\n",
"unsafe_code": false,
"doc_to_choice": "def doc_to_choice(doc):\n idx = doc[\"sentence\"].index(\"_\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"sentence\"][:idx] + opt for opt in options]\n",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "default",
"split": null,
"process_docs": null,
"fewshot_indices": null,
"samples": null,
"doc_to_text": "<function doc_to_text at 0x73f093d3ca40>",
"doc_to_choice": "<function doc_to_choice at 0x73f093d3d080>",
"doc_to_target": "<function doc_to_target at 0x73f093d3ce00>",
"gen_prefix": null,
"fewshot_delimiter": "\n\n",
"target_delimiter": " "
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": true,
"doc_to_decontamination_query": "sentence",
"metadata": {
"version": 1.0,
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
}
}
},
"versions": {
"arc_challenge": 1.0,
"arc_easy": 1.0,
"hellaswag": 1.0,
"kmmlu": 2.0,
"kmmlu_accounting": 2.0,
"kmmlu_agricultural_sciences": 2.0,
"kmmlu_applied_science": 2.0,
"kmmlu_aviation_engineering_and_maintenance": 2.0,
"kmmlu_biology": 2.0,
"kmmlu_chemical_engineering": 2.0,
"kmmlu_chemistry": 2.0,
"kmmlu_civil_engineering": 2.0,
"kmmlu_computer_science": 2.0,
"kmmlu_construction": 2.0,
"kmmlu_criminal_law": 2.0,
"kmmlu_ecology": 2.0,
"kmmlu_economics": 2.0,
"kmmlu_education": 2.0,
"kmmlu_electrical_engineering": 2.0,
"kmmlu_electronics_engineering": 2.0,
"kmmlu_energy_management": 2.0,
"kmmlu_environmental_science": 2.0,
"kmmlu_fashion": 2.0,
"kmmlu_food_processing": 2.0,
"kmmlu_gas_technology_and_engineering": 2.0,
"kmmlu_geomatics": 2.0,
"kmmlu_health": 2.0,
"kmmlu_humss": 2.0,
"kmmlu_industrial_engineer": 2.0,
"kmmlu_information_technology": 2.0,
"kmmlu_interior_architecture_and_design": 2.0,
"kmmlu_korean_history": 2.0,
"kmmlu_law": 2.0,
"kmmlu_machine_design_and_manufacturing": 2.0,
"kmmlu_management": 2.0,
"kmmlu_maritime_engineering": 2.0,
"kmmlu_marketing": 2.0,
"kmmlu_materials_engineering": 2.0,
"kmmlu_math": 2.0,
"kmmlu_mechanical_engineering": 2.0,
"kmmlu_nondestructive_testing": 2.0,
"kmmlu_other": 2.0,
"kmmlu_patent": 2.0,
"kmmlu_political_science_and_sociology": 2.0,
"kmmlu_psychology": 2.0,
"kmmlu_public_safety": 2.0,
"kmmlu_railway_and_automotive_engineering": 2.0,
"kmmlu_real_estate": 2.0,
"kmmlu_refrigerating_machinery": 2.0,
"kmmlu_social_welfare": 2.0,
"kmmlu_stem": 2.0,
"kmmlu_taxation": 2.0,
"kmmlu_telecommunications_and_wireless_technology": 2.0,
"kobest_boolq": 1.0,
"kobest_copa": 1.0,
"kobest_hellaswag": 1.0,
"mmlu": 2,
"mmlu_abstract_algebra": 1.0,
"mmlu_anatomy": 1.0,
"mmlu_astronomy": 1.0,
"mmlu_business_ethics": 1.0,
"mmlu_clinical_knowledge": 1.0,
"mmlu_college_biology": 1.0,
"mmlu_college_chemistry": 1.0,
"mmlu_college_computer_science": 1.0,
"mmlu_college_mathematics": 1.0,
"mmlu_college_medicine": 1.0,
"mmlu_college_physics": 1.0,
"mmlu_computer_security": 1.0,
"mmlu_conceptual_physics": 1.0,
"mmlu_econometrics": 1.0,
"mmlu_electrical_engineering": 1.0,
"mmlu_elementary_mathematics": 1.0,
"mmlu_formal_logic": 1.0,
"mmlu_global_facts": 1.0,
"mmlu_high_school_biology": 1.0,
"mmlu_high_school_chemistry": 1.0,
"mmlu_high_school_computer_science": 1.0,
"mmlu_high_school_european_history": 1.0,
"mmlu_high_school_geography": 1.0,
"mmlu_high_school_government_and_politics": 1.0,
"mmlu_high_school_macroeconomics": 1.0,
"mmlu_high_school_mathematics": 1.0,
"mmlu_high_school_microeconomics": 1.0,
"mmlu_high_school_physics": 1.0,
"mmlu_high_school_psychology": 1.0,
"mmlu_high_school_statistics": 1.0,
"mmlu_high_school_us_history": 1.0,
"mmlu_high_school_world_history": 1.0,
"mmlu_human_aging": 1.0,
"mmlu_human_sexuality": 1.0,
"mmlu_humanities": 2,
"mmlu_international_law": 1.0,
"mmlu_jurisprudence": 1.0,
"mmlu_logical_fallacies": 1.0,
"mmlu_machine_learning": 1.0,
"mmlu_management": 1.0,
"mmlu_marketing": 1.0,
"mmlu_medical_genetics": 1.0,
"mmlu_miscellaneous": 1.0,
"mmlu_moral_disputes": 1.0,
"mmlu_moral_scenarios": 1.0,
"mmlu_nutrition": 1.0,
"mmlu_other": 2,
"mmlu_philosophy": 1.0,
"mmlu_prehistory": 1.0,
"mmlu_professional_accounting": 1.0,
"mmlu_professional_law": 1.0,
"mmlu_professional_medicine": 1.0,
"mmlu_professional_psychology": 1.0,
"mmlu_public_relations": 1.0,
"mmlu_security_studies": 1.0,
"mmlu_social_sciences": 2,
"mmlu_sociology": 1.0,
"mmlu_stem": 2,
"mmlu_us_foreign_policy": 1.0,
"mmlu_virology": 1.0,
"mmlu_world_religions": 1.0,
"winogrande": 1.0
},
"n-shot": {
"arc_challenge": 0,
"arc_easy": 0,
"hellaswag": 0,
"kmmlu_accounting": 0,
"kmmlu_agricultural_sciences": 0,
"kmmlu_aviation_engineering_and_maintenance": 0,
"kmmlu_biology": 0,
"kmmlu_chemical_engineering": 0,
"kmmlu_chemistry": 0,
"kmmlu_civil_engineering": 0,
"kmmlu_computer_science": 0,
"kmmlu_construction": 0,
"kmmlu_criminal_law": 0,
"kmmlu_ecology": 0,
"kmmlu_economics": 0,
"kmmlu_education": 0,
"kmmlu_electrical_engineering": 0,
"kmmlu_electronics_engineering": 0,
"kmmlu_energy_management": 0,
"kmmlu_environmental_science": 0,
"kmmlu_fashion": 0,
"kmmlu_food_processing": 0,
"kmmlu_gas_technology_and_engineering": 0,
"kmmlu_geomatics": 0,
"kmmlu_health": 0,
"kmmlu_industrial_engineer": 0,
"kmmlu_information_technology": 0,
"kmmlu_interior_architecture_and_design": 0,
"kmmlu_korean_history": 0,
"kmmlu_law": 0,
"kmmlu_machine_design_and_manufacturing": 0,
"kmmlu_management": 0,
"kmmlu_maritime_engineering": 0,
"kmmlu_marketing": 0,
"kmmlu_materials_engineering": 0,
"kmmlu_math": 0,
"kmmlu_mechanical_engineering": 0,
"kmmlu_nondestructive_testing": 0,
"kmmlu_patent": 0,
"kmmlu_political_science_and_sociology": 0,
"kmmlu_psychology": 0,
"kmmlu_public_safety": 0,
"kmmlu_railway_and_automotive_engineering": 0,
"kmmlu_real_estate": 0,
"kmmlu_refrigerating_machinery": 0,
"kmmlu_social_welfare": 0,
"kmmlu_taxation": 0,
"kmmlu_telecommunications_and_wireless_technology": 0,
"kobest_boolq": 0,
"kobest_copa": 0,
"kobest_hellaswag": 0,
"mmlu_abstract_algebra": 0,
"mmlu_anatomy": 0,
"mmlu_astronomy": 0,
"mmlu_business_ethics": 0,
"mmlu_clinical_knowledge": 0,
"mmlu_college_biology": 0,
"mmlu_college_chemistry": 0,
"mmlu_college_computer_science": 0,
"mmlu_college_mathematics": 0,
"mmlu_college_medicine": 0,
"mmlu_college_physics": 0,
"mmlu_computer_security": 0,
"mmlu_conceptual_physics": 0,
"mmlu_econometrics": 0,
"mmlu_electrical_engineering": 0,
"mmlu_elementary_mathematics": 0,
"mmlu_formal_logic": 0,
"mmlu_global_facts": 0,
"mmlu_high_school_biology": 0,
"mmlu_high_school_chemistry": 0,
"mmlu_high_school_computer_science": 0,
"mmlu_high_school_european_history": 0,
"mmlu_high_school_geography": 0,
"mmlu_high_school_government_and_politics": 0,
"mmlu_high_school_macroeconomics": 0,
"mmlu_high_school_mathematics": 0,
"mmlu_high_school_microeconomics": 0,
"mmlu_high_school_physics": 0,
"mmlu_high_school_psychology": 0,
"mmlu_high_school_statistics": 0,
"mmlu_high_school_us_history": 0,
"mmlu_high_school_world_history": 0,
"mmlu_human_aging": 0,
"mmlu_human_sexuality": 0,
"mmlu_international_law": 0,
"mmlu_jurisprudence": 0,
"mmlu_logical_fallacies": 0,
"mmlu_machine_learning": 0,
"mmlu_management": 0,
"mmlu_marketing": 0,
"mmlu_medical_genetics": 0,
"mmlu_miscellaneous": 0,
"mmlu_moral_disputes": 0,
"mmlu_moral_scenarios": 0,
"mmlu_nutrition": 0,
"mmlu_philosophy": 0,
"mmlu_prehistory": 0,
"mmlu_professional_accounting": 0,
"mmlu_professional_law": 0,
"mmlu_professional_medicine": 0,
"mmlu_professional_psychology": 0,
"mmlu_public_relations": 0,
"mmlu_security_studies": 0,
"mmlu_sociology": 0,
"mmlu_us_foreign_policy": 0,
"mmlu_virology": 0,
"mmlu_world_religions": 0,
"winogrande": 0
},
"higher_is_better": {
"arc_challenge": {
"acc": true,
"acc_norm": true
},
"arc_easy": {
"acc": true,
"acc_norm": true
},
"hellaswag": {
"acc": true,
"acc_norm": true
},
"kmmlu": {
"acc": true
},
"kmmlu_accounting": {
"acc": true
},
"kmmlu_agricultural_sciences": {
"acc": true
},
"kmmlu_applied_science": {
"acc": true
},
"kmmlu_aviation_engineering_and_maintenance": {
"acc": true
},
"kmmlu_biology": {
"acc": true
},
"kmmlu_chemical_engineering": {
"acc": true
},
"kmmlu_chemistry": {
"acc": true
},
"kmmlu_civil_engineering": {
"acc": true
},
"kmmlu_computer_science": {
"acc": true
},
"kmmlu_construction": {
"acc": true
},
"kmmlu_criminal_law": {
"acc": true
},
"kmmlu_ecology": {
"acc": true
},
"kmmlu_economics": {
"acc": true
},
"kmmlu_education": {
"acc": true
},
"kmmlu_electrical_engineering": {
"acc": true
},
"kmmlu_electronics_engineering": {
"acc": true
},
"kmmlu_energy_management": {
"acc": true
},
"kmmlu_environmental_science": {
"acc": true
},
"kmmlu_fashion": {
"acc": true
},
"kmmlu_food_processing": {
"acc": true
},
"kmmlu_gas_technology_and_engineering": {
"acc": true
},
"kmmlu_geomatics": {
"acc": true
},
"kmmlu_health": {
"acc": true
},
"kmmlu_humss": {
"acc": true
},
"kmmlu_industrial_engineer": {
"acc": true
},
"kmmlu_information_technology": {
"acc": true
},
"kmmlu_interior_architecture_and_design": {
"acc": true
},
"kmmlu_korean_history": {
"acc": true
},
"kmmlu_law": {
"acc": true
},
"kmmlu_machine_design_and_manufacturing": {
"acc": true
},
"kmmlu_management": {
"acc": true
},
"kmmlu_maritime_engineering": {
"acc": true
},
"kmmlu_marketing": {
"acc": true
},
"kmmlu_materials_engineering": {
"acc": true
},
"kmmlu_math": {
"acc": true
},
"kmmlu_mechanical_engineering": {
"acc": true
},
"kmmlu_nondestructive_testing": {
"acc": true
},
"kmmlu_other": {
"acc": true
},
"kmmlu_patent": {
"acc": true
},
"kmmlu_political_science_and_sociology": {
"acc": true
},
"kmmlu_psychology": {
"acc": true
},
"kmmlu_public_safety": {
"acc": true
},
"kmmlu_railway_and_automotive_engineering": {
"acc": true
},
"kmmlu_real_estate": {
"acc": true
},
"kmmlu_refrigerating_machinery": {
"acc": true
},
"kmmlu_social_welfare": {
"acc": true
},
"kmmlu_stem": {
"acc": true
},
"kmmlu_taxation": {
"acc": true
},
"kmmlu_telecommunications_and_wireless_technology": {
"acc": true
},
"kobest_boolq": {
"acc": true,
"f1": true
},
"kobest_copa": {
"acc": true,
"f1": true
},
"kobest_hellaswag": {
"acc": true,
"acc_norm": true,
"f1": true
},
"mmlu": {
"acc": true
},
"mmlu_abstract_algebra": {
"acc": true
},
"mmlu_anatomy": {
"acc": true
},
"mmlu_astronomy": {
"acc": true
},
"mmlu_business_ethics": {
"acc": true
},
"mmlu_clinical_knowledge": {
"acc": true
},
"mmlu_college_biology": {
"acc": true
},
"mmlu_college_chemistry": {
"acc": true
},
"mmlu_college_computer_science": {
"acc": true
},
"mmlu_college_mathematics": {
"acc": true
},
"mmlu_college_medicine": {
"acc": true
},
"mmlu_college_physics": {
"acc": true
},
"mmlu_computer_security": {
"acc": true
},
"mmlu_conceptual_physics": {
"acc": true
},
"mmlu_econometrics": {
"acc": true
},
"mmlu_electrical_engineering": {
"acc": true
},
"mmlu_elementary_mathematics": {
"acc": true
},
"mmlu_formal_logic": {
"acc": true
},
"mmlu_global_facts": {
"acc": true
},
"mmlu_high_school_biology": {
"acc": true
},
"mmlu_high_school_chemistry": {
"acc": true
},
"mmlu_high_school_computer_science": {
"acc": true
},
"mmlu_high_school_european_history": {
"acc": true
},
"mmlu_high_school_geography": {
"acc": true
},
"mmlu_high_school_government_and_politics": {
"acc": true
},
"mmlu_high_school_macroeconomics": {
"acc": true
},
"mmlu_high_school_mathematics": {
"acc": true
},
"mmlu_high_school_microeconomics": {
"acc": true
},
"mmlu_high_school_physics": {
"acc": true
},
"mmlu_high_school_psychology": {
"acc": true
},
"mmlu_high_school_statistics": {
"acc": true
},
"mmlu_high_school_us_history": {
"acc": true
},
"mmlu_high_school_world_history": {
"acc": true
},
"mmlu_human_aging": {
"acc": true
},
"mmlu_human_sexuality": {
"acc": true
},
"mmlu_humanities": {
"acc": true
},
"mmlu_international_law": {
"acc": true
},
"mmlu_jurisprudence": {
"acc": true
},
"mmlu_logical_fallacies": {
"acc": true
},
"mmlu_machine_learning": {
"acc": true
},
"mmlu_management": {
"acc": true
},
"mmlu_marketing": {
"acc": true
},
"mmlu_medical_genetics": {
"acc": true
},
"mmlu_miscellaneous": {
"acc": true
},
"mmlu_moral_disputes": {
"acc": true
},
"mmlu_moral_scenarios": {
"acc": true
},
"mmlu_nutrition": {
"acc": true
},
"mmlu_other": {
"acc": true
},
"mmlu_philosophy": {
"acc": true
},
"mmlu_prehistory": {
"acc": true
},
"mmlu_professional_accounting": {
"acc": true
},
"mmlu_professional_law": {
"acc": true
},
"mmlu_professional_medicine": {
"acc": true
},
"mmlu_professional_psychology": {
"acc": true
},
"mmlu_public_relations": {
"acc": true
},
"mmlu_security_studies": {
"acc": true
},
"mmlu_social_sciences": {
"acc": true
},
"mmlu_sociology": {
"acc": true
},
"mmlu_stem": {
"acc": true
},
"mmlu_us_foreign_policy": {
"acc": true
},
"mmlu_virology": {
"acc": true
},
"mmlu_world_religions": {
"acc": true
},
"winogrande": {
"acc": true
}
},
"n-samples": {
"kobest_hellaswag": {
"original": 500,
"effective": 400
},
"kobest_copa": {
"original": 1000,
"effective": 400
},
"kobest_boolq": {
"original": 1404,
"effective": 400
},
"kmmlu_biology": {
"original": 1000,
"effective": 400
},
"kmmlu_chemical_engineering": {
"original": 1000,
"effective": 400
},
"kmmlu_chemistry": {
"original": 600,
"effective": 400
},
"kmmlu_civil_engineering": {
"original": 1000,
"effective": 400
},
"kmmlu_computer_science": {
"original": 1000,
"effective": 400
},
"kmmlu_ecology": {
"original": 1000,
"effective": 400
},
"kmmlu_electrical_engineering": {
"original": 1000,
"effective": 400
},
"kmmlu_information_technology": {
"original": 1000,
"effective": 400
},
"kmmlu_materials_engineering": {
"original": 1000,
"effective": 400
},
"kmmlu_math": {
"original": 300,
"effective": 300
},
"kmmlu_mechanical_engineering": {
"original": 1000,
"effective": 400
},
"kmmlu_agricultural_sciences": {
"original": 1000,
"effective": 400
},
"kmmlu_construction": {
"original": 1000,
"effective": 400
},
"kmmlu_fashion": {
"original": 1000,
"effective": 400
},
"kmmlu_food_processing": {
"original": 1000,
"effective": 400
},
"kmmlu_health": {
"original": 100,
"effective": 100
},
"kmmlu_interior_architecture_and_design": {
"original": 1000,
"effective": 400
},
"kmmlu_marketing": {
"original": 1000,
"effective": 400
},
"kmmlu_patent": {
"original": 100,
"effective": 100
},
"kmmlu_public_safety": {
"original": 1000,
"effective": 400
},
"kmmlu_real_estate": {
"original": 200,
"effective": 200
},
"kmmlu_refrigerating_machinery": {
"original": 1000,
"effective": 400
},
"kmmlu_aviation_engineering_and_maintenance": {
"original": 1000,
"effective": 400
},
"kmmlu_electronics_engineering": {
"original": 1000,
"effective": 400
},
"kmmlu_energy_management": {
"original": 1000,
"effective": 400
},
"kmmlu_environmental_science": {
"original": 1000,
"effective": 400
},
"kmmlu_gas_technology_and_engineering": {
"original": 1000,
"effective": 400
},
"kmmlu_geomatics": {
"original": 1000,
"effective": 400
},
"kmmlu_industrial_engineer": {
"original": 1000,
"effective": 400
},
"kmmlu_machine_design_and_manufacturing": {
"original": 1000,
"effective": 400
},
"kmmlu_maritime_engineering": {
"original": 600,
"effective": 400
},
"kmmlu_nondestructive_testing": {
"original": 1000,
"effective": 400
},
"kmmlu_railway_and_automotive_engineering": {
"original": 1000,
"effective": 400
},
"kmmlu_telecommunications_and_wireless_technology": {
"original": 1000,
"effective": 400
},
"kmmlu_accounting": {
"original": 100,
"effective": 100
},
"kmmlu_criminal_law": {
"original": 200,
"effective": 200
},
"kmmlu_economics": {
"original": 130,
"effective": 130
},
"kmmlu_education": {
"original": 100,
"effective": 100
},
"kmmlu_korean_history": {
"original": 100,
"effective": 100
},
"kmmlu_law": {
"original": 1000,
"effective": 400
},
"kmmlu_management": {
"original": 1000,
"effective": 400
},
"kmmlu_political_science_and_sociology": {
"original": 300,
"effective": 300
},
"kmmlu_psychology": {
"original": 1000,
"effective": 400
},
"kmmlu_social_welfare": {
"original": 1000,
"effective": 400
},
"kmmlu_taxation": {
"original": 200,
"effective": 200
},
"winogrande": {
"original": 1267,
"effective": 400
},
"arc_challenge": {
"original": 1172,
"effective": 400
},
"arc_easy": {
"original": 2376,
"effective": 400
},
"hellaswag": {
"original": 10042,
"effective": 400
},
"mmlu_abstract_algebra": {
"original": 100,
"effective": 100
},
"mmlu_anatomy": {
"original": 135,
"effective": 135
},
"mmlu_astronomy": {
"original": 152,
"effective": 152
},
"mmlu_college_biology": {
"original": 144,
"effective": 144
},
"mmlu_college_chemistry": {
"original": 100,
"effective": 100
},
"mmlu_college_computer_science": {
"original": 100,
"effective": 100
},
"mmlu_college_mathematics": {
"original": 100,
"effective": 100
},
"mmlu_college_physics": {
"original": 102,
"effective": 102
},
"mmlu_computer_security": {
"original": 100,
"effective": 100
},
"mmlu_conceptual_physics": {
"original": 235,
"effective": 235
},
"mmlu_electrical_engineering": {
"original": 145,
"effective": 145
},
"mmlu_elementary_mathematics": {
"original": 378,
"effective": 378
},
"mmlu_high_school_biology": {
"original": 310,
"effective": 310
},
"mmlu_high_school_chemistry": {
"original": 203,
"effective": 203
},
"mmlu_high_school_computer_science": {
"original": 100,
"effective": 100
},
"mmlu_high_school_mathematics": {
"original": 270,
"effective": 270
},
"mmlu_high_school_physics": {
"original": 151,
"effective": 151
},
"mmlu_high_school_statistics": {
"original": 216,
"effective": 216
},
"mmlu_machine_learning": {
"original": 112,
"effective": 112
},
"mmlu_business_ethics": {
"original": 100,
"effective": 100
},
"mmlu_clinical_knowledge": {
"original": 265,
"effective": 265
},
"mmlu_college_medicine": {
"original": 173,
"effective": 173
},
"mmlu_global_facts": {
"original": 100,
"effective": 100
},
"mmlu_human_aging": {
"original": 223,
"effective": 223
},
"mmlu_management": {
"original": 103,
"effective": 103
},
"mmlu_marketing": {
"original": 234,
"effective": 234
},
"mmlu_medical_genetics": {
"original": 100,
"effective": 100
},
"mmlu_miscellaneous": {
"original": 783,
"effective": 400
},
"mmlu_nutrition": {
"original": 306,
"effective": 306
},
"mmlu_professional_accounting": {
"original": 282,
"effective": 282
},
"mmlu_professional_medicine": {
"original": 272,
"effective": 272
},
"mmlu_virology": {
"original": 166,
"effective": 166
},
"mmlu_econometrics": {
"original": 114,
"effective": 114
},
"mmlu_high_school_geography": {
"original": 198,
"effective": 198
},
"mmlu_high_school_government_and_politics": {
"original": 193,
"effective": 193
},
"mmlu_high_school_macroeconomics": {
"original": 390,
"effective": 390
},
"mmlu_high_school_microeconomics": {
"original": 238,
"effective": 238
},
"mmlu_high_school_psychology": {
"original": 545,
"effective": 400
},
"mmlu_human_sexuality": {
"original": 131,
"effective": 131
},
"mmlu_professional_psychology": {
"original": 612,
"effective": 400
},
"mmlu_public_relations": {
"original": 110,
"effective": 110
},
"mmlu_security_studies": {
"original": 245,
"effective": 245
},
"mmlu_sociology": {
"original": 201,
"effective": 201
},
"mmlu_us_foreign_policy": {
"original": 100,
"effective": 100
},
"mmlu_formal_logic": {
"original": 126,
"effective": 126
},
"mmlu_high_school_european_history": {
"original": 165,
"effective": 165
},
"mmlu_high_school_us_history": {
"original": 204,
"effective": 204
},
"mmlu_high_school_world_history": {
"original": 237,
"effective": 237
},
"mmlu_international_law": {
"original": 121,
"effective": 121
},
"mmlu_jurisprudence": {
"original": 108,
"effective": 108
},
"mmlu_logical_fallacies": {
"original": 163,
"effective": 163
},
"mmlu_moral_disputes": {
"original": 346,
"effective": 346
},
"mmlu_moral_scenarios": {
"original": 895,
"effective": 400
},
"mmlu_philosophy": {
"original": 311,
"effective": 311
},
"mmlu_prehistory": {
"original": 324,
"effective": 324
},
"mmlu_professional_law": {
"original": 1534,
"effective": 400
},
"mmlu_world_religions": {
"original": 171,
"effective": 171
}
},
"config": {
"model": "hf",
"model_args": {
"pretrained": "unsloth/Qwen3-4B-Base",
"trust_remote_code": true
},
"model_num_parameters": 4022468096,
"model_dtype": "torch.bfloat16",
"model_revision": "main",
"model_sha": "0573b584bc6b32adc84bb9c91bf9b71bea71fc40",
"batch_size": "12",
"batch_sizes": [],
"device": "cuda:0",
"use_cache": null,
"limit": 400.0,
"bootstrap_iters": 100000,
"gen_kwargs": {},
"random_seed": 0,
"numpy_seed": 1234,
"torch_seed": 1234,
"fewshot_seed": 1234
},
"git_hash": "0ce43af",
"date": 1775962695.520946,
"pretty_env_info": "PyTorch version: 2.9.0+cu128\nIs debug build: False\nCUDA used to build PyTorch: 12.8\nROCM used to build PyTorch: N/A\n\nOS: Ubuntu 22.04.5 LTS (x86_64)\nGCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0\nClang version: Could not collect\nCMake version: version 4.1.0\nLibc version: glibc-2.35\n\nPython version: 3.11.14 | packaged by conda-forge | (main, Oct 13 2025, 14:09:32) [GCC 14.3.0] (64-bit runtime)\nPython platform: Linux-6.8.0-90-generic-x86_64-with-glibc2.35\nIs CUDA available: True\nCUDA runtime version: 12.8.93\nCUDA_MODULE_LOADING set to: \nGPU models and configuration: GPU 0: NVIDIA RTX PRO 6000 Blackwell Workstation Edition\nNvidia driver version: 590.48.01\ncuDNN version: Probably one of the following:\n/usr/lib/x86_64-linux-gnu/libcudnn.so.9.8.0\n/usr/lib/x86_64-linux-gnu/libcudnn_adv.so.9.8.0\n/usr/lib/x86_64-linux-gnu/libcudnn_cnn.so.9.8.0\n/usr/lib/x86_64-linux-gnu/libcudnn_engines_precompiled.so.9.8.0\n/usr/lib/x86_64-linux-gnu/libcudnn_engines_runtime_compiled.so.9.8.0\n/usr/lib/x86_64-linux-gnu/libcudnn_graph.so.9.8.0\n/usr/lib/x86_64-linux-gnu/libcudnn_heuristic.so.9.8.0\n/usr/lib/x86_64-linux-gnu/libcudnn_ops.so.9.8.0\nIs XPU available: False\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nAddress sizes: 43 bits physical, 48 bits virtual\nByte Order: Little Endian\nCPU(s): 192\nOn-line CPU(s) list: 0-191\nVendor ID: AuthenticAMD\nModel name: AMD EPYC 7642 48-Core Processor\nCPU family: 23\nModel: 49\nThread(s) per core: 2\nCore(s) per socket: 48\nSocket(s): 2\nStepping: 0\nFrequency boost: enabled\nCPU max MHz: 2300.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 4600.15\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid aperfmperf rapl pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_llc mwaitx cpb cat_l3 cdp_l3 hw_pstate ssbd mba ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr rdpru wbnoinvd amd_ppin arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif v_spec_ctrl umip rdpid overflow_recov succor smca sev sev_es ibpb_exit_to_user\nVirtualization: AMD-V\nL1d cache: 3 MiB (96 instances)\nL1i cache: 3 MiB (96 instances)\nL2 cache: 48 MiB (96 instances)\nL3 cache: 512 MiB (32 instances)\nNUMA node(s): 2\nNUMA node0 CPU(s): 0-47,96-143\nNUMA node1 CPU(s): 48-95,144-191\nVulnerability Gather data sampling: Not affected\nVulnerability Itlb multihit: Not affected\nVulnerability L1tf: Not affected\nVulnerability Mds: Not affected\nVulnerability Meltdown: Not affected\nVulnerability Mmio stale data: Not affected\nVulnerability Reg file data sampling: Not affected\nVulnerability Retbleed: Mitigation; untrained return thunk; SMT enabled with STIBP protection\nVulnerability Spec rs
"transformers_version": "5.5.3",
"lm_eval_version": "0.4.11",
"upper_git_hash": null,
"tokenizer_pad_token": [
"<|vision_pad|>",
"151654"
],
"tokenizer_eos_token": [
"<|endoftext|>",
"151643"
],
"tokenizer_bos_token": [
null,
"None"
],
"eot_token_id": 151643,
"max_length": 32768,
"task_hashes": {},
"model_source": "hf",
"model_name": "unsloth/Qwen3-4B-Base",
"model_name_sanitized": "unsloth__Qwen3-4B-Base",
"system_instruction": null,
"system_instruction_sha": null,
"fewshot_as_multiturn": null,
"chat_template": null,
"chat_template_sha": null,
"total_evaluation_time_seconds": "573.7631184216589"
}
}
}