Sean Cho commited on
Commit
81e9f42
1 Parent(s): 0feedb7
choco9966/Llama-2-7b-instruct-tuning/result_2023-10-19 08:44:42.json DELETED
@@ -1,444 +0,0 @@
1
- {
2
- "results": {
3
- "harness|ko_arc_challenge|25": {
4
- "acc": 0.29692832764505117,
5
- "acc_stderr": 0.013352025976725222,
6
- "acc_norm": 0.33447098976109213,
7
- "acc_norm_stderr": 0.01378746032244138
8
- },
9
- "harness|ko_hellaswag|10": {
10
- "acc": 0.345947022505477,
11
- "acc_stderr": 0.004747038768172532,
12
- "acc_norm": 0.4251145190201155,
13
- "acc_norm_stderr": 0.004933500261683597
14
- },
15
- "harness|ko_mmlu_world_religions|5": {
16
- "acc": 0.4269005847953216,
17
- "acc_stderr": 0.03793620616529916,
18
- "acc_norm": 0.4269005847953216,
19
- "acc_norm_stderr": 0.03793620616529916
20
- },
21
- "harness|ko_mmlu_management|5": {
22
- "acc": 0.34951456310679613,
23
- "acc_stderr": 0.047211885060971716,
24
- "acc_norm": 0.34951456310679613,
25
- "acc_norm_stderr": 0.047211885060971716
26
- },
27
- "harness|ko_mmlu_miscellaneous|5": {
28
- "acc": 0.3895274584929757,
29
- "acc_stderr": 0.017438082556264594,
30
- "acc_norm": 0.3895274584929757,
31
- "acc_norm_stderr": 0.017438082556264594
32
- },
33
- "harness|ko_mmlu_anatomy|5": {
34
- "acc": 0.31851851851851853,
35
- "acc_stderr": 0.040247784019771124,
36
- "acc_norm": 0.31851851851851853,
37
- "acc_norm_stderr": 0.040247784019771124
38
- },
39
- "harness|ko_mmlu_abstract_algebra|5": {
40
- "acc": 0.38,
41
- "acc_stderr": 0.048783173121456316,
42
- "acc_norm": 0.38,
43
- "acc_norm_stderr": 0.048783173121456316
44
- },
45
- "harness|ko_mmlu_conceptual_physics|5": {
46
- "acc": 0.3659574468085106,
47
- "acc_stderr": 0.03148955829745529,
48
- "acc_norm": 0.3659574468085106,
49
- "acc_norm_stderr": 0.03148955829745529
50
- },
51
- "harness|ko_mmlu_virology|5": {
52
- "acc": 0.29518072289156627,
53
- "acc_stderr": 0.03550920185689629,
54
- "acc_norm": 0.29518072289156627,
55
- "acc_norm_stderr": 0.03550920185689629
56
- },
57
- "harness|ko_mmlu_philosophy|5": {
58
- "acc": 0.3858520900321543,
59
- "acc_stderr": 0.02764814959975147,
60
- "acc_norm": 0.3858520900321543,
61
- "acc_norm_stderr": 0.02764814959975147
62
- },
63
- "harness|ko_mmlu_human_aging|5": {
64
- "acc": 0.34977578475336324,
65
- "acc_stderr": 0.03200736719484503,
66
- "acc_norm": 0.34977578475336324,
67
- "acc_norm_stderr": 0.03200736719484503
68
- },
69
- "harness|ko_mmlu_human_sexuality|5": {
70
- "acc": 0.45038167938931295,
71
- "acc_stderr": 0.04363643698524779,
72
- "acc_norm": 0.45038167938931295,
73
- "acc_norm_stderr": 0.04363643698524779
74
- },
75
- "harness|ko_mmlu_medical_genetics|5": {
76
- "acc": 0.35,
77
- "acc_stderr": 0.04793724854411021,
78
- "acc_norm": 0.35,
79
- "acc_norm_stderr": 0.04793724854411021
80
- },
81
- "harness|ko_mmlu_high_school_geography|5": {
82
- "acc": 0.35353535353535354,
83
- "acc_stderr": 0.03406086723547153,
84
- "acc_norm": 0.35353535353535354,
85
- "acc_norm_stderr": 0.03406086723547153
86
- },
87
- "harness|ko_mmlu_electrical_engineering|5": {
88
- "acc": 0.3724137931034483,
89
- "acc_stderr": 0.040287315329475604,
90
- "acc_norm": 0.3724137931034483,
91
- "acc_norm_stderr": 0.040287315329475604
92
- },
93
- "harness|ko_mmlu_college_physics|5": {
94
- "acc": 0.21568627450980393,
95
- "acc_stderr": 0.04092563958237654,
96
- "acc_norm": 0.21568627450980393,
97
- "acc_norm_stderr": 0.04092563958237654
98
- },
99
- "harness|ko_mmlu_high_school_microeconomics|5": {
100
- "acc": 0.33613445378151263,
101
- "acc_stderr": 0.030684737115135377,
102
- "acc_norm": 0.33613445378151263,
103
- "acc_norm_stderr": 0.030684737115135377
104
- },
105
- "harness|ko_mmlu_high_school_macroeconomics|5": {
106
- "acc": 0.31025641025641026,
107
- "acc_stderr": 0.023454674889404288,
108
- "acc_norm": 0.31025641025641026,
109
- "acc_norm_stderr": 0.023454674889404288
110
- },
111
- "harness|ko_mmlu_computer_security|5": {
112
- "acc": 0.37,
113
- "acc_stderr": 0.04852365870939099,
114
- "acc_norm": 0.37,
115
- "acc_norm_stderr": 0.04852365870939099
116
- },
117
- "harness|ko_mmlu_global_facts|5": {
118
- "acc": 0.37,
119
- "acc_stderr": 0.048523658709391,
120
- "acc_norm": 0.37,
121
- "acc_norm_stderr": 0.048523658709391
122
- },
123
- "harness|ko_mmlu_jurisprudence|5": {
124
- "acc": 0.3888888888888889,
125
- "acc_stderr": 0.047128212574267705,
126
- "acc_norm": 0.3888888888888889,
127
- "acc_norm_stderr": 0.047128212574267705
128
- },
129
- "harness|ko_mmlu_high_school_chemistry|5": {
130
- "acc": 0.27586206896551724,
131
- "acc_stderr": 0.0314471258167824,
132
- "acc_norm": 0.27586206896551724,
133
- "acc_norm_stderr": 0.0314471258167824
134
- },
135
- "harness|ko_mmlu_high_school_biology|5": {
136
- "acc": 0.3580645161290323,
137
- "acc_stderr": 0.027273890594300642,
138
- "acc_norm": 0.3580645161290323,
139
- "acc_norm_stderr": 0.027273890594300642
140
- },
141
- "harness|ko_mmlu_marketing|5": {
142
- "acc": 0.5128205128205128,
143
- "acc_stderr": 0.032745319388423504,
144
- "acc_norm": 0.5128205128205128,
145
- "acc_norm_stderr": 0.032745319388423504
146
- },
147
- "harness|ko_mmlu_clinical_knowledge|5": {
148
- "acc": 0.32452830188679244,
149
- "acc_stderr": 0.028815615713432115,
150
- "acc_norm": 0.32452830188679244,
151
- "acc_norm_stderr": 0.028815615713432115
152
- },
153
- "harness|ko_mmlu_public_relations|5": {
154
- "acc": 0.39090909090909093,
155
- "acc_stderr": 0.04673752333670237,
156
- "acc_norm": 0.39090909090909093,
157
- "acc_norm_stderr": 0.04673752333670237
158
- },
159
- "harness|ko_mmlu_high_school_mathematics|5": {
160
- "acc": 0.29259259259259257,
161
- "acc_stderr": 0.027738969632176095,
162
- "acc_norm": 0.29259259259259257,
163
- "acc_norm_stderr": 0.027738969632176095
164
- },
165
- "harness|ko_mmlu_high_school_physics|5": {
166
- "acc": 0.26490066225165565,
167
- "acc_stderr": 0.03603038545360384,
168
- "acc_norm": 0.26490066225165565,
169
- "acc_norm_stderr": 0.03603038545360384
170
- },
171
- "harness|ko_mmlu_sociology|5": {
172
- "acc": 0.48258706467661694,
173
- "acc_stderr": 0.03533389234739245,
174
- "acc_norm": 0.48258706467661694,
175
- "acc_norm_stderr": 0.03533389234739245
176
- },
177
- "harness|ko_mmlu_college_medicine|5": {
178
- "acc": 0.31213872832369943,
179
- "acc_stderr": 0.035331333893236574,
180
- "acc_norm": 0.31213872832369943,
181
- "acc_norm_stderr": 0.035331333893236574
182
- },
183
- "harness|ko_mmlu_elementary_mathematics|5": {
184
- "acc": 0.30158730158730157,
185
- "acc_stderr": 0.0236369759961018,
186
- "acc_norm": 0.30158730158730157,
187
- "acc_norm_stderr": 0.0236369759961018
188
- },
189
- "harness|ko_mmlu_college_biology|5": {
190
- "acc": 0.3055555555555556,
191
- "acc_stderr": 0.03852084696008534,
192
- "acc_norm": 0.3055555555555556,
193
- "acc_norm_stderr": 0.03852084696008534
194
- },
195
- "harness|ko_mmlu_college_chemistry|5": {
196
- "acc": 0.31,
197
- "acc_stderr": 0.04648231987117316,
198
- "acc_norm": 0.31,
199
- "acc_norm_stderr": 0.04648231987117316
200
- },
201
- "harness|ko_mmlu_us_foreign_policy|5": {
202
- "acc": 0.44,
203
- "acc_stderr": 0.04988876515698589,
204
- "acc_norm": 0.44,
205
- "acc_norm_stderr": 0.04988876515698589
206
- },
207
- "harness|ko_mmlu_moral_disputes|5": {
208
- "acc": 0.3439306358381503,
209
- "acc_stderr": 0.025574123786546648,
210
- "acc_norm": 0.3439306358381503,
211
- "acc_norm_stderr": 0.025574123786546648
212
- },
213
- "harness|ko_mmlu_logical_fallacies|5": {
214
- "acc": 0.34355828220858897,
215
- "acc_stderr": 0.03731133519673893,
216
- "acc_norm": 0.34355828220858897,
217
- "acc_norm_stderr": 0.03731133519673893
218
- },
219
- "harness|ko_mmlu_prehistory|5": {
220
- "acc": 0.38271604938271603,
221
- "acc_stderr": 0.027044538138402616,
222
- "acc_norm": 0.38271604938271603,
223
- "acc_norm_stderr": 0.027044538138402616
224
- },
225
- "harness|ko_mmlu_college_mathematics|5": {
226
- "acc": 0.28,
227
- "acc_stderr": 0.04512608598542129,
228
- "acc_norm": 0.28,
229
- "acc_norm_stderr": 0.04512608598542129
230
- },
231
- "harness|ko_mmlu_high_school_government_and_politics|5": {
232
- "acc": 0.43005181347150256,
233
- "acc_stderr": 0.03572954333144808,
234
- "acc_norm": 0.43005181347150256,
235
- "acc_norm_stderr": 0.03572954333144808
236
- },
237
- "harness|ko_mmlu_econometrics|5": {
238
- "acc": 0.2719298245614035,
239
- "acc_stderr": 0.041857744240220575,
240
- "acc_norm": 0.2719298245614035,
241
- "acc_norm_stderr": 0.041857744240220575
242
- },
243
- "harness|ko_mmlu_high_school_psychology|5": {
244
- "acc": 0.3412844036697248,
245
- "acc_stderr": 0.020328612816592432,
246
- "acc_norm": 0.3412844036697248,
247
- "acc_norm_stderr": 0.020328612816592432
248
- },
249
- "harness|ko_mmlu_formal_logic|5": {
250
- "acc": 0.30158730158730157,
251
- "acc_stderr": 0.04104947269903394,
252
- "acc_norm": 0.30158730158730157,
253
- "acc_norm_stderr": 0.04104947269903394
254
- },
255
- "harness|ko_mmlu_nutrition|5": {
256
- "acc": 0.39215686274509803,
257
- "acc_stderr": 0.027956046165424516,
258
- "acc_norm": 0.39215686274509803,
259
- "acc_norm_stderr": 0.027956046165424516
260
- },
261
- "harness|ko_mmlu_business_ethics|5": {
262
- "acc": 0.46,
263
- "acc_stderr": 0.05009082659620332,
264
- "acc_norm": 0.46,
265
- "acc_norm_stderr": 0.05009082659620332
266
- },
267
- "harness|ko_mmlu_international_law|5": {
268
- "acc": 0.49586776859504134,
269
- "acc_stderr": 0.04564198767432754,
270
- "acc_norm": 0.49586776859504134,
271
- "acc_norm_stderr": 0.04564198767432754
272
- },
273
- "harness|ko_mmlu_astronomy|5": {
274
- "acc": 0.35526315789473684,
275
- "acc_stderr": 0.03894734487013315,
276
- "acc_norm": 0.35526315789473684,
277
- "acc_norm_stderr": 0.03894734487013315
278
- },
279
- "harness|ko_mmlu_professional_psychology|5": {
280
- "acc": 0.29411764705882354,
281
- "acc_stderr": 0.018433427649401896,
282
- "acc_norm": 0.29411764705882354,
283
- "acc_norm_stderr": 0.018433427649401896
284
- },
285
- "harness|ko_mmlu_professional_accounting|5": {
286
- "acc": 0.2765957446808511,
287
- "acc_stderr": 0.026684564340460987,
288
- "acc_norm": 0.2765957446808511,
289
- "acc_norm_stderr": 0.026684564340460987
290
- },
291
- "harness|ko_mmlu_machine_learning|5": {
292
- "acc": 0.24107142857142858,
293
- "acc_stderr": 0.04059867246952687,
294
- "acc_norm": 0.24107142857142858,
295
- "acc_norm_stderr": 0.04059867246952687
296
- },
297
- "harness|ko_mmlu_high_school_statistics|5": {
298
- "acc": 0.4074074074074074,
299
- "acc_stderr": 0.03350991604696043,
300
- "acc_norm": 0.4074074074074074,
301
- "acc_norm_stderr": 0.03350991604696043
302
- },
303
- "harness|ko_mmlu_moral_scenarios|5": {
304
- "acc": 0.24804469273743016,
305
- "acc_stderr": 0.01444415780826145,
306
- "acc_norm": 0.24804469273743016,
307
- "acc_norm_stderr": 0.01444415780826145
308
- },
309
- "harness|ko_mmlu_college_computer_science|5": {
310
- "acc": 0.34,
311
- "acc_stderr": 0.04760952285695235,
312
- "acc_norm": 0.34,
313
- "acc_norm_stderr": 0.04760952285695235
314
- },
315
- "harness|ko_mmlu_high_school_computer_science|5": {
316
- "acc": 0.34,
317
- "acc_stderr": 0.04760952285695236,
318
- "acc_norm": 0.34,
319
- "acc_norm_stderr": 0.04760952285695236
320
- },
321
- "harness|ko_mmlu_professional_medicine|5": {
322
- "acc": 0.3088235294117647,
323
- "acc_stderr": 0.028064998167040094,
324
- "acc_norm": 0.3088235294117647,
325
- "acc_norm_stderr": 0.028064998167040094
326
- },
327
- "harness|ko_mmlu_security_studies|5": {
328
- "acc": 0.47346938775510206,
329
- "acc_stderr": 0.03196412734523272,
330
- "acc_norm": 0.47346938775510206,
331
- "acc_norm_stderr": 0.03196412734523272
332
- },
333
- "harness|ko_mmlu_high_school_world_history|5": {
334
- "acc": 0.35443037974683544,
335
- "acc_stderr": 0.031137304297185798,
336
- "acc_norm": 0.35443037974683544,
337
- "acc_norm_stderr": 0.031137304297185798
338
- },
339
- "harness|ko_mmlu_professional_law|5": {
340
- "acc": 0.2627118644067797,
341
- "acc_stderr": 0.011240545514995669,
342
- "acc_norm": 0.2627118644067797,
343
- "acc_norm_stderr": 0.011240545514995669
344
- },
345
- "harness|ko_mmlu_high_school_us_history|5": {
346
- "acc": 0.31862745098039214,
347
- "acc_stderr": 0.032702871814820816,
348
- "acc_norm": 0.31862745098039214,
349
- "acc_norm_stderr": 0.032702871814820816
350
- },
351
- "harness|ko_mmlu_high_school_european_history|5": {
352
- "acc": 0.38181818181818183,
353
- "acc_stderr": 0.03793713171165634,
354
- "acc_norm": 0.38181818181818183,
355
- "acc_norm_stderr": 0.03793713171165634
356
- },
357
- "harness|ko_truthfulqa_mc|0": {
358
- "mc1": 0.33047735618115054,
359
- "mc1_stderr": 0.016466769613698293,
360
- "mc2": 0.5139753799906011,
361
- "mc2_stderr": 0.016082624616035393
362
- },
363
- "harness|ko_commongen_v2|2": {
364
- "acc": 0.29225352112676056,
365
- "acc_stderr": 0.015590281423747496,
366
- "acc_norm": 0.32511737089201875,
367
- "acc_norm_stderr": 0.016057185777207585
368
- }
369
- },
370
- "versions": {
371
- "all": 0,
372
- "harness|ko_arc_challenge|25": 0,
373
- "harness|ko_hellaswag|10": 0,
374
- "harness|ko_mmlu_world_religions|5": 1,
375
- "harness|ko_mmlu_management|5": 1,
376
- "harness|ko_mmlu_miscellaneous|5": 1,
377
- "harness|ko_mmlu_anatomy|5": 1,
378
- "harness|ko_mmlu_abstract_algebra|5": 1,
379
- "harness|ko_mmlu_conceptual_physics|5": 1,
380
- "harness|ko_mmlu_virology|5": 1,
381
- "harness|ko_mmlu_philosophy|5": 1,
382
- "harness|ko_mmlu_human_aging|5": 1,
383
- "harness|ko_mmlu_human_sexuality|5": 1,
384
- "harness|ko_mmlu_medical_genetics|5": 1,
385
- "harness|ko_mmlu_high_school_geography|5": 1,
386
- "harness|ko_mmlu_electrical_engineering|5": 1,
387
- "harness|ko_mmlu_college_physics|5": 1,
388
- "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
- "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
- "harness|ko_mmlu_computer_security|5": 1,
391
- "harness|ko_mmlu_global_facts|5": 1,
392
- "harness|ko_mmlu_jurisprudence|5": 1,
393
- "harness|ko_mmlu_high_school_chemistry|5": 1,
394
- "harness|ko_mmlu_high_school_biology|5": 1,
395
- "harness|ko_mmlu_marketing|5": 1,
396
- "harness|ko_mmlu_clinical_knowledge|5": 1,
397
- "harness|ko_mmlu_public_relations|5": 1,
398
- "harness|ko_mmlu_high_school_mathematics|5": 1,
399
- "harness|ko_mmlu_high_school_physics|5": 1,
400
- "harness|ko_mmlu_sociology|5": 1,
401
- "harness|ko_mmlu_college_medicine|5": 1,
402
- "harness|ko_mmlu_elementary_mathematics|5": 1,
403
- "harness|ko_mmlu_college_biology|5": 1,
404
- "harness|ko_mmlu_college_chemistry|5": 1,
405
- "harness|ko_mmlu_us_foreign_policy|5": 1,
406
- "harness|ko_mmlu_moral_disputes|5": 1,
407
- "harness|ko_mmlu_logical_fallacies|5": 1,
408
- "harness|ko_mmlu_prehistory|5": 1,
409
- "harness|ko_mmlu_college_mathematics|5": 1,
410
- "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
- "harness|ko_mmlu_econometrics|5": 1,
412
- "harness|ko_mmlu_high_school_psychology|5": 1,
413
- "harness|ko_mmlu_formal_logic|5": 1,
414
- "harness|ko_mmlu_nutrition|5": 1,
415
- "harness|ko_mmlu_business_ethics|5": 1,
416
- "harness|ko_mmlu_international_law|5": 1,
417
- "harness|ko_mmlu_astronomy|5": 1,
418
- "harness|ko_mmlu_professional_psychology|5": 1,
419
- "harness|ko_mmlu_professional_accounting|5": 1,
420
- "harness|ko_mmlu_machine_learning|5": 1,
421
- "harness|ko_mmlu_high_school_statistics|5": 1,
422
- "harness|ko_mmlu_moral_scenarios|5": 1,
423
- "harness|ko_mmlu_college_computer_science|5": 1,
424
- "harness|ko_mmlu_high_school_computer_science|5": 1,
425
- "harness|ko_mmlu_professional_medicine|5": 1,
426
- "harness|ko_mmlu_security_studies|5": 1,
427
- "harness|ko_mmlu_high_school_world_history|5": 1,
428
- "harness|ko_mmlu_professional_law|5": 1,
429
- "harness|ko_mmlu_high_school_us_history|5": 1,
430
- "harness|ko_mmlu_high_school_european_history|5": 1,
431
- "harness|ko_truthfulqa_mc|0": 0,
432
- "harness|ko_commongen_v2|2": 1
433
- },
434
- "config_general": {
435
- "model_name": "choco9966/Llama-2-7b-instruct-tuning",
436
- "model_sha": "0914768714fca5e74eef736b357d9f82ccc9e089",
437
- "model_dtype": "torch.float16",
438
- "lighteval_sha": "",
439
- "num_few_shot_default": 0,
440
- "num_fewshot_seeds": 1,
441
- "override_batch_size": 1,
442
- "max_samples": null
443
- }
444
- }