{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.29266211604095566, "acc_stderr": 0.013295916103619413, "acc_norm": 0.31399317406143346, "acc_norm_stderr": 0.013562691224726291 }, "harness|ko_hellaswag|10": { "acc": 0.37004580760804623, "acc_stderr": 0.00481829899101255, "acc_norm": 0.47470623381796456, "acc_norm_stderr": 0.004983392650570958 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.17543859649122806, "acc_stderr": 0.029170885500727665, "acc_norm": 0.17543859649122806, "acc_norm_stderr": 0.029170885500727665 }, "harness|ko_mmlu_management|5": { "acc": 0.24271844660194175, "acc_stderr": 0.04245022486384495, "acc_norm": 0.24271844660194175, "acc_norm_stderr": 0.04245022486384495 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.20434227330779056, "acc_stderr": 0.014419123980931906, "acc_norm": 0.20434227330779056, "acc_norm_stderr": 0.014419123980931906 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.22962962962962963, "acc_stderr": 0.036333844140734636, "acc_norm": 0.22962962962962963, "acc_norm_stderr": 0.036333844140734636 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.18, "acc_stderr": 0.03861229196653694, "acc_norm": 0.18, "acc_norm_stderr": 0.03861229196653694 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.22127659574468084, "acc_stderr": 0.027136349602424063, "acc_norm": 0.22127659574468084, "acc_norm_stderr": 0.027136349602424063 }, "harness|ko_mmlu_virology|5": { "acc": 0.22289156626506024, "acc_stderr": 0.03240004825594688, "acc_norm": 0.22289156626506024, "acc_norm_stderr": 0.03240004825594688 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.24115755627009647, "acc_stderr": 0.024296594034763426, "acc_norm": 0.24115755627009647, "acc_norm_stderr": 0.024296594034763426 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.11659192825112108, "acc_stderr": 0.021539639816244464, "acc_norm": 0.11659192825112108, "acc_norm_stderr": 0.021539639816244464 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.2824427480916031, "acc_stderr": 0.03948406125768361, "acc_norm": 0.2824427480916031, "acc_norm_stderr": 0.03948406125768361 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.25, "acc_stderr": 0.04351941398892446, "acc_norm": 0.25, "acc_norm_stderr": 0.04351941398892446 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.3434343434343434, "acc_stderr": 0.03383201223244441, "acc_norm": 0.3434343434343434, "acc_norm_stderr": 0.03383201223244441 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.18620689655172415, "acc_stderr": 0.03243946159004616, "acc_norm": 0.18620689655172415, "acc_norm_stderr": 0.03243946159004616 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.30392156862745096, "acc_stderr": 0.04576665403207763, "acc_norm": 0.30392156862745096, "acc_norm_stderr": 0.04576665403207763 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.2857142857142857, "acc_stderr": 0.02934457250063435, "acc_norm": 0.2857142857142857, "acc_norm_stderr": 0.02934457250063435 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.358974358974359, "acc_stderr": 0.024321738484602364, "acc_norm": 0.358974358974359, "acc_norm_stderr": 0.024321738484602364 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.19, "acc_stderr": 0.03942772444036625, "acc_norm": 0.19, "acc_norm_stderr": 0.03942772444036625 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.18, "acc_stderr": 0.038612291966536955, "acc_norm": 0.18, "acc_norm_stderr": 0.038612291966536955 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.21296296296296297, "acc_stderr": 0.03957835471980981, "acc_norm": 0.21296296296296297, "acc_norm_stderr": 0.03957835471980981 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.2413793103448276, "acc_stderr": 0.030108330718011625, "acc_norm": 0.2413793103448276, "acc_norm_stderr": 0.030108330718011625 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.3258064516129032, "acc_stderr": 0.0266620105785671, "acc_norm": 0.3258064516129032, "acc_norm_stderr": 0.0266620105785671 }, "harness|ko_mmlu_marketing|5": { "acc": 0.19658119658119658, "acc_stderr": 0.02603538609895129, "acc_norm": 0.19658119658119658, "acc_norm_stderr": 0.02603538609895129 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.27547169811320754, "acc_stderr": 0.02749566368372406, "acc_norm": 0.27547169811320754, "acc_norm_stderr": 0.02749566368372406 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.23636363636363636, "acc_stderr": 0.04069306319721376, "acc_norm": 0.23636363636363636, "acc_norm_stderr": 0.04069306319721376 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.26666666666666666, "acc_stderr": 0.026962424325073835, "acc_norm": 0.26666666666666666, "acc_norm_stderr": 0.026962424325073835 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.3509933774834437, "acc_stderr": 0.03896981964257374, "acc_norm": 0.3509933774834437, "acc_norm_stderr": 0.03896981964257374 }, "harness|ko_mmlu_sociology|5": { "acc": 0.26865671641791045, "acc_stderr": 0.03134328358208954, "acc_norm": 0.26865671641791045, "acc_norm_stderr": 0.03134328358208954 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.3352601156069364, "acc_stderr": 0.03599586301247078, "acc_norm": 0.3352601156069364, "acc_norm_stderr": 0.03599586301247078 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.25925925925925924, "acc_stderr": 0.02256989707491842, "acc_norm": 0.25925925925925924, "acc_norm_stderr": 0.02256989707491842 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.2638888888888889, "acc_stderr": 0.03685651095897532, "acc_norm": 0.2638888888888889, "acc_norm_stderr": 0.03685651095897532 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.27, "acc_stderr": 0.04461960433384741, "acc_norm": 0.27, "acc_norm_stderr": 0.04461960433384741 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.26878612716763006, "acc_stderr": 0.023868003262500114, "acc_norm": 0.26878612716763006, "acc_norm_stderr": 0.023868003262500114 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.2392638036809816, "acc_stderr": 0.03351953879521271, "acc_norm": 0.2392638036809816, "acc_norm_stderr": 0.03351953879521271 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.23148148148148148, "acc_stderr": 0.02346842983245115, "acc_norm": 0.23148148148148148, "acc_norm_stderr": 0.02346842983245115 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.36787564766839376, "acc_stderr": 0.034801756684660366, "acc_norm": 0.36787564766839376, "acc_norm_stderr": 0.034801756684660366 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2543859649122807, "acc_stderr": 0.040969851398436695, "acc_norm": 0.2543859649122807, "acc_norm_stderr": 0.040969851398436695 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.3431192660550459, "acc_stderr": 0.02035477773608604, "acc_norm": 0.3431192660550459, "acc_norm_stderr": 0.02035477773608604 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.3412698412698413, "acc_stderr": 0.04240799327574924, "acc_norm": 0.3412698412698413, "acc_norm_stderr": 0.04240799327574924 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.2908496732026144, "acc_stderr": 0.026004800363952113, "acc_norm": 0.2908496732026144, "acc_norm_stderr": 0.026004800363952113 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.26, "acc_stderr": 0.04408440022768078, "acc_norm": 0.26, "acc_norm_stderr": 0.04408440022768078 }, "harness|ko_mmlu_international_law|5": { "acc": 0.17355371900826447, "acc_stderr": 0.03457272836917669, "acc_norm": 0.17355371900826447, "acc_norm_stderr": 0.03457272836917669 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.3355263157894737, "acc_stderr": 0.038424985593952694, "acc_norm": 0.3355263157894737, "acc_norm_stderr": 0.038424985593952694 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.2173202614379085, "acc_stderr": 0.0166848209291486, "acc_norm": 0.2173202614379085, "acc_norm_stderr": 0.0166848209291486 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.20921985815602837, "acc_stderr": 0.02426476943998849, "acc_norm": 0.20921985815602837, "acc_norm_stderr": 0.02426476943998849 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.19642857142857142, "acc_stderr": 0.03770970049347018, "acc_norm": 0.19642857142857142, "acc_norm_stderr": 0.03770970049347018 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.4722222222222222, "acc_stderr": 0.0340470532865388, "acc_norm": 0.4722222222222222, "acc_norm_stderr": 0.0340470532865388 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.27262569832402234, "acc_stderr": 0.014893391735249608, "acc_norm": 0.27262569832402234, "acc_norm_stderr": 0.014893391735249608 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.34, "acc_stderr": 0.04760952285695235, "acc_norm": 0.34, "acc_norm_stderr": 0.04760952285695235 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.21, "acc_stderr": 0.04093601807403325, "acc_norm": 0.21, "acc_norm_stderr": 0.04093601807403325 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.4485294117647059, "acc_stderr": 0.030211479609121593, "acc_norm": 0.4485294117647059, "acc_norm_stderr": 0.030211479609121593 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.3836734693877551, "acc_stderr": 0.03113088039623593, "acc_norm": 0.3836734693877551, "acc_norm_stderr": 0.03113088039623593 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.25316455696202533, "acc_stderr": 0.0283046579430353, "acc_norm": 0.25316455696202533, "acc_norm_stderr": 0.0283046579430353 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.24641460234680573, "acc_stderr": 0.011005971399927235, "acc_norm": 0.24641460234680573, "acc_norm_stderr": 0.011005971399927235 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.25, "acc_stderr": 0.03039153369274154, "acc_norm": 0.25, "acc_norm_stderr": 0.03039153369274154 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.2727272727272727, "acc_stderr": 0.0347769116216366, "acc_norm": 0.2727272727272727, "acc_norm_stderr": 0.0347769116216366 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.24112607099143207, "mc1_stderr": 0.014974827279752332, "mc2": 0.39158327266747156, "mc2_stderr": 0.014622481693781006 }, "harness|ko_commongen_v2|2": { "acc": 0.3685446009389671, "acc_stderr": 0.01653680430615456, "acc_norm": 0.4706572769953052, "acc_norm_stderr": 0.017110239257076235 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "amphora/polyglot-5.8B-CoT-e1", "model_sha": "e8f4cb1d884cf4d67e3e8afc0aab09c62a0d68c6", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }