open-ko-llm-bot commited on
Commit
8dc6202
1 Parent(s): c9a312f

Add results for 2023-10-01 02:38:38

Browse files
yeen214/test_llama2_7b/result_2023-10-01 02:38:38.json CHANGED
@@ -13,10 +13,10 @@
13
  "acc_norm_stderr": 0.004910588449330016
14
  },
15
  "harness|ko_mmlu_world_religions|5": {
16
- "acc": 0.4853801169590643,
17
- "acc_stderr": 0.038331852752130205,
18
- "acc_norm": 0.4853801169590643,
19
- "acc_norm_stderr": 0.038331852752130205
20
  },
21
  "harness|ko_mmlu_management|5": {
22
  "acc": 0.3106796116504854,
@@ -25,10 +25,10 @@
25
  "acc_norm_stderr": 0.045821241601615506
26
  },
27
  "harness|ko_mmlu_miscellaneous|5": {
28
- "acc": 0.37037037037037035,
29
- "acc_stderr": 0.017268607560005773,
30
- "acc_norm": 0.37037037037037035,
31
- "acc_norm_stderr": 0.017268607560005773
32
  },
33
  "harness|ko_mmlu_anatomy|5": {
34
  "acc": 0.3037037037037037,
@@ -91,10 +91,10 @@
91
  "acc_norm_stderr": 0.039966295748767186
92
  },
93
  "harness|ko_mmlu_college_physics|5": {
94
- "acc": 0.18627450980392157,
95
- "acc_stderr": 0.038739587141493524,
96
- "acc_norm": 0.18627450980392157,
97
- "acc_norm_stderr": 0.038739587141493524
98
  },
99
  "harness|ko_mmlu_high_school_microeconomics|5": {
100
  "acc": 0.3235294117647059,
@@ -169,10 +169,10 @@
169
  "acc_norm_stderr": 0.033367670865679766
170
  },
171
  "harness|ko_mmlu_sociology|5": {
172
- "acc": 0.44776119402985076,
173
- "acc_stderr": 0.035161847729521675,
174
- "acc_norm": 0.44776119402985076,
175
- "acc_norm_stderr": 0.035161847729521675
176
  },
177
  "harness|ko_mmlu_college_medicine|5": {
178
  "acc": 0.26011560693641617,
@@ -181,10 +181,10 @@
181
  "acc_norm_stderr": 0.03345036916788991
182
  },
183
  "harness|ko_mmlu_elementary_mathematics|5": {
184
- "acc": 0.2724867724867725,
185
- "acc_stderr": 0.022930973071633356,
186
- "acc_norm": 0.2724867724867725,
187
- "acc_norm_stderr": 0.022930973071633356
188
  },
189
  "harness|ko_mmlu_college_biology|5": {
190
  "acc": 0.2847222222222222,
@@ -229,10 +229,10 @@
229
  "acc_norm_stderr": 0.046056618647183814
230
  },
231
  "harness|ko_mmlu_high_school_government_and_politics|5": {
232
- "acc": 0.35233160621761656,
233
- "acc_stderr": 0.034474782864143544,
234
- "acc_norm": 0.35233160621761656,
235
- "acc_norm_stderr": 0.034474782864143544
236
  },
237
  "harness|ko_mmlu_econometrics|5": {
238
  "acc": 0.23684210526315788,
@@ -283,16 +283,16 @@
283
  "acc_norm_stderr": 0.018690850273595284
284
  },
285
  "harness|ko_mmlu_professional_accounting|5": {
286
- "acc": 0.28368794326241137,
287
- "acc_stderr": 0.02689170942834396,
288
- "acc_norm": 0.28368794326241137,
289
- "acc_norm_stderr": 0.02689170942834396
290
  },
291
  "harness|ko_mmlu_machine_learning|5": {
292
- "acc": 0.30357142857142855,
293
- "acc_stderr": 0.04364226155841044,
294
- "acc_norm": 0.30357142857142855,
295
- "acc_norm_stderr": 0.04364226155841044
296
  },
297
  "harness|ko_mmlu_high_school_statistics|5": {
298
  "acc": 0.37962962962962965,
@@ -307,10 +307,10 @@
307
  "acc_norm_stderr": 0.01435591196476786
308
  },
309
  "harness|ko_mmlu_college_computer_science|5": {
310
- "acc": 0.31,
311
- "acc_stderr": 0.04648231987117316,
312
- "acc_norm": 0.31,
313
- "acc_norm_stderr": 0.04648231987117316
314
  },
315
  "harness|ko_mmlu_high_school_computer_science|5": {
316
  "acc": 0.33,
 
13
  "acc_norm_stderr": 0.004910588449330016
14
  },
15
  "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.47953216374269003,
17
+ "acc_stderr": 0.038316105328219316,
18
+ "acc_norm": 0.47953216374269003,
19
+ "acc_norm_stderr": 0.038316105328219316
20
  },
21
  "harness|ko_mmlu_management|5": {
22
  "acc": 0.3106796116504854,
 
25
  "acc_norm_stderr": 0.045821241601615506
26
  },
27
  "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.36909323116219667,
29
+ "acc_stderr": 0.017256283109124613,
30
+ "acc_norm": 0.36909323116219667,
31
+ "acc_norm_stderr": 0.017256283109124613
32
  },
33
  "harness|ko_mmlu_anatomy|5": {
34
  "acc": 0.3037037037037037,
 
91
  "acc_norm_stderr": 0.039966295748767186
92
  },
93
  "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.16666666666666666,
95
+ "acc_stderr": 0.03708284662416545,
96
+ "acc_norm": 0.16666666666666666,
97
+ "acc_norm_stderr": 0.03708284662416545
98
  },
99
  "harness|ko_mmlu_high_school_microeconomics|5": {
100
  "acc": 0.3235294117647059,
 
169
  "acc_norm_stderr": 0.033367670865679766
170
  },
171
  "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.4577114427860697,
173
+ "acc_stderr": 0.03522865864099597,
174
+ "acc_norm": 0.4577114427860697,
175
+ "acc_norm_stderr": 0.03522865864099597
176
  },
177
  "harness|ko_mmlu_college_medicine|5": {
178
  "acc": 0.26011560693641617,
 
181
  "acc_norm_stderr": 0.03345036916788991
182
  },
183
  "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.2751322751322751,
185
+ "acc_stderr": 0.02300008685906864,
186
+ "acc_norm": 0.2751322751322751,
187
+ "acc_norm_stderr": 0.02300008685906864
188
  },
189
  "harness|ko_mmlu_college_biology|5": {
190
  "acc": 0.2847222222222222,
 
229
  "acc_norm_stderr": 0.046056618647183814
230
  },
231
  "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.35751295336787564,
233
+ "acc_stderr": 0.03458816042181007,
234
+ "acc_norm": 0.35751295336787564,
235
+ "acc_norm_stderr": 0.03458816042181007
236
  },
237
  "harness|ko_mmlu_econometrics|5": {
238
  "acc": 0.23684210526315788,
 
283
  "acc_norm_stderr": 0.018690850273595284
284
  },
285
  "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.2872340425531915,
287
+ "acc_stderr": 0.026992199173064356,
288
+ "acc_norm": 0.2872340425531915,
289
+ "acc_norm_stderr": 0.026992199173064356
290
  },
291
  "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.3125,
293
+ "acc_stderr": 0.043994650575715215,
294
+ "acc_norm": 0.3125,
295
+ "acc_norm_stderr": 0.043994650575715215
296
  },
297
  "harness|ko_mmlu_high_school_statistics|5": {
298
  "acc": 0.37962962962962965,
 
307
  "acc_norm_stderr": 0.01435591196476786
308
  },
309
  "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.3,
311
+ "acc_stderr": 0.046056618647183814,
312
+ "acc_norm": 0.3,
313
+ "acc_norm_stderr": 0.046056618647183814
314
  },
315
  "harness|ko_mmlu_high_school_computer_science|5": {
316
  "acc": 0.33,