{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.2226962457337884, "acc_stderr": 0.012158314774829948, "acc_norm": 0.2627986348122867, "acc_norm_stderr": 0.012862523175351331 }, "harness|ko_hellaswag|10": { "acc": 0.2726548496315475, "acc_stderr": 0.004444146875436292, "acc_norm": 0.29635530770762797, "acc_norm_stderr": 0.004557163175885563 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.2982456140350877, "acc_stderr": 0.03508771929824561, "acc_norm": 0.2982456140350877, "acc_norm_stderr": 0.03508771929824561 }, "harness|ko_mmlu_management|5": { "acc": 0.3106796116504854, "acc_stderr": 0.04582124160161549, "acc_norm": 0.3106796116504854, "acc_norm_stderr": 0.04582124160161549 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.2848020434227331, "acc_stderr": 0.016139174096522553, "acc_norm": 0.2848020434227331, "acc_norm_stderr": 0.016139174096522553 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.25925925925925924, "acc_stderr": 0.03785714465066654, "acc_norm": 0.25925925925925924, "acc_norm_stderr": 0.03785714465066654 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.3276595744680851, "acc_stderr": 0.030683020843231008, "acc_norm": 0.3276595744680851, "acc_norm_stderr": 0.030683020843231008 }, "harness|ko_mmlu_virology|5": { "acc": 0.2891566265060241, "acc_stderr": 0.03529486801511115, "acc_norm": 0.2891566265060241, "acc_norm_stderr": 0.03529486801511115 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.3440514469453376, "acc_stderr": 0.026981478043648026, "acc_norm": 0.3440514469453376, "acc_norm_stderr": 0.026981478043648026 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.34977578475336324, "acc_stderr": 0.03200736719484503, "acc_norm": 0.34977578475336324, "acc_norm_stderr": 0.03200736719484503 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.2366412213740458, "acc_stderr": 0.037276735755969195, "acc_norm": 0.2366412213740458, "acc_norm_stderr": 0.037276735755969195 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.22727272727272727, "acc_stderr": 0.029857515673386417, "acc_norm": 0.22727272727272727, "acc_norm_stderr": 0.029857515673386417 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.31724137931034485, "acc_stderr": 0.038783523721386215, "acc_norm": 0.31724137931034485, "acc_norm_stderr": 0.038783523721386215 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.17647058823529413, "acc_stderr": 0.03793281185307811, "acc_norm": 0.17647058823529413, "acc_norm_stderr": 0.03793281185307811 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.2689075630252101, "acc_stderr": 0.028801392193631276, "acc_norm": 0.2689075630252101, "acc_norm_stderr": 0.028801392193631276 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.2512820512820513, "acc_stderr": 0.021992016662370568, "acc_norm": 0.2512820512820513, "acc_norm_stderr": 0.021992016662370568 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.29, "acc_stderr": 0.045604802157206845, "acc_norm": 0.29, "acc_norm_stderr": 0.045604802157206845 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.33, "acc_stderr": 0.04725815626252604, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252604 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.3611111111111111, "acc_stderr": 0.04643454608906275, "acc_norm": 0.3611111111111111, "acc_norm_stderr": 0.04643454608906275 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.26108374384236455, "acc_stderr": 0.030903796952114454, "acc_norm": 0.26108374384236455, "acc_norm_stderr": 0.030903796952114454 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.2709677419354839, "acc_stderr": 0.02528441611490016, "acc_norm": 0.2709677419354839, "acc_norm_stderr": 0.02528441611490016 }, "harness|ko_mmlu_marketing|5": { "acc": 0.3418803418803419, "acc_stderr": 0.03107502852650775, "acc_norm": 0.3418803418803419, "acc_norm_stderr": 0.03107502852650775 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.2943396226415094, "acc_stderr": 0.028049186315695248, "acc_norm": 0.2943396226415094, "acc_norm_stderr": 0.028049186315695248 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.3181818181818182, "acc_stderr": 0.044612721759105065, "acc_norm": 0.3181818181818182, "acc_norm_stderr": 0.044612721759105065 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.2962962962962963, "acc_stderr": 0.027840811495871927, "acc_norm": 0.2962962962962963, "acc_norm_stderr": 0.027840811495871927 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.2582781456953642, "acc_stderr": 0.035737053147634576, "acc_norm": 0.2582781456953642, "acc_norm_stderr": 0.035737053147634576 }, "harness|ko_mmlu_sociology|5": { "acc": 0.38308457711442784, "acc_stderr": 0.034375193373382504, "acc_norm": 0.38308457711442784, "acc_norm_stderr": 0.034375193373382504 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.24855491329479767, "acc_stderr": 0.03295304696818318, "acc_norm": 0.24855491329479767, "acc_norm_stderr": 0.03295304696818318 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.2724867724867725, "acc_stderr": 0.022930973071633345, "acc_norm": 0.2724867724867725, "acc_norm_stderr": 0.022930973071633345 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.25, "acc_stderr": 0.03621034121889507, "acc_norm": 0.25, "acc_norm_stderr": 0.03621034121889507 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.22, "acc_stderr": 0.04163331998932269, "acc_norm": 0.22, "acc_norm_stderr": 0.04163331998932269 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.35, "acc_stderr": 0.0479372485441102, "acc_norm": 0.35, "acc_norm_stderr": 0.0479372485441102 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.315028901734104, "acc_stderr": 0.025009313790069692, "acc_norm": 0.315028901734104, "acc_norm_stderr": 0.025009313790069692 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.3006134969325153, "acc_stderr": 0.03602511318806771, "acc_norm": 0.3006134969325153, "acc_norm_stderr": 0.03602511318806771 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.30864197530864196, "acc_stderr": 0.02570264026060375, "acc_norm": 0.30864197530864196, "acc_norm_stderr": 0.02570264026060375 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.28, "acc_stderr": 0.04512608598542127, "acc_norm": 0.28, "acc_norm_stderr": 0.04512608598542127 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.3160621761658031, "acc_stderr": 0.03355397369686173, "acc_norm": 0.3160621761658031, "acc_norm_stderr": 0.03355397369686173 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2719298245614035, "acc_stderr": 0.041857744240220575, "acc_norm": 0.2719298245614035, "acc_norm_stderr": 0.041857744240220575 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.23853211009174313, "acc_stderr": 0.01827257581023186, "acc_norm": 0.23853211009174313, "acc_norm_stderr": 0.01827257581023186 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.2857142857142857, "acc_stderr": 0.04040610178208841, "acc_norm": 0.2857142857142857, "acc_norm_stderr": 0.04040610178208841 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.2908496732026144, "acc_stderr": 0.026004800363952113, "acc_norm": 0.2908496732026144, "acc_norm_stderr": 0.026004800363952113 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.23, "acc_stderr": 0.04229525846816508, "acc_norm": 0.23, "acc_norm_stderr": 0.04229525846816508 }, "harness|ko_mmlu_international_law|5": { "acc": 0.512396694214876, "acc_stderr": 0.04562951548180765, "acc_norm": 0.512396694214876, "acc_norm_stderr": 0.04562951548180765 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.2236842105263158, "acc_stderr": 0.03391160934343604, "acc_norm": 0.2236842105263158, "acc_norm_stderr": 0.03391160934343604 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.2696078431372549, "acc_stderr": 0.017952449196987866, "acc_norm": 0.2696078431372549, "acc_norm_stderr": 0.017952449196987866 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.26595744680851063, "acc_stderr": 0.026358065698880585, "acc_norm": 0.26595744680851063, "acc_norm_stderr": 0.026358065698880585 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.29464285714285715, "acc_stderr": 0.043270409325787296, "acc_norm": 0.29464285714285715, "acc_norm_stderr": 0.043270409325787296 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.19907407407407407, "acc_stderr": 0.027232298462690218, "acc_norm": 0.19907407407407407, "acc_norm_stderr": 0.027232298462690218 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.2424581005586592, "acc_stderr": 0.01433352205921789, "acc_norm": 0.2424581005586592, "acc_norm_stderr": 0.01433352205921789 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.23, "acc_stderr": 0.04229525846816506, "acc_norm": 0.23, "acc_norm_stderr": 0.04229525846816506 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.26, "acc_stderr": 0.04408440022768079, "acc_norm": 0.26, "acc_norm_stderr": 0.04408440022768079 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.22058823529411764, "acc_stderr": 0.02518778666022727, "acc_norm": 0.22058823529411764, "acc_norm_stderr": 0.02518778666022727 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.20408163265306123, "acc_stderr": 0.025801283475090496, "acc_norm": 0.20408163265306123, "acc_norm_stderr": 0.025801283475090496 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.3037974683544304, "acc_stderr": 0.029936696387138598, "acc_norm": 0.3037974683544304, "acc_norm_stderr": 0.029936696387138598 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.24837027379400262, "acc_stderr": 0.011035212598034494, "acc_norm": 0.24837027379400262, "acc_norm_stderr": 0.011035212598034494 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.24019607843137256, "acc_stderr": 0.02998373305591362, "acc_norm": 0.24019607843137256, "acc_norm_stderr": 0.02998373305591362 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.24242424242424243, "acc_stderr": 0.033464098810559534, "acc_norm": 0.24242424242424243, "acc_norm_stderr": 0.033464098810559534 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2594859241126071, "mc1_stderr": 0.015345409485557966, "mc2": 0.43443146146429873, "mc2_stderr": 0.01580310882533787 }, "harness|ko_commongen_v2|2": { "acc": 0.22183098591549297, "acc_stderr": 0.014242403694199452, "acc_norm": 0.3626760563380282, "acc_norm_stderr": 0.016480666823965092 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "jb723/llama2-ko-7B-model", "model_sha": "03d23910fa0f9b0542ce7634cbcd36983321f55a", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }