teknium commited on
Commit
59ea029
1 Parent(s): 9206f0f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -0
README.md CHANGED
@@ -69,6 +69,32 @@ GPT-4All Benchmark Set
69
  Average: 0.679
70
  ```
71
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
72
  TruthfulQA:
73
  ```
74
  hf-causal-experimental (pretrained=teknium/OpenHermes-7B,dtype=float16), limit: None, provide_description: False, num_fewshot: 0, batch_size: 8
 
69
  Average: 0.679
70
  ```
71
 
72
+ BigBench:
73
+ ```
74
+ | Task |Version| Metric |Value | |Stderr|
75
+ |------------------------------------------------|------:|---------------------|-----:|---|-----:|
76
+ |bigbench_causal_judgement | 0|multiple_choice_grade|0.5000|± |0.0364|
77
+ |bigbench_date_understanding | 0|multiple_choice_grade|0.5908|± |0.0256|
78
+ |bigbench_disambiguation_qa | 0|multiple_choice_grade|0.3023|± |0.0286|
79
+ |bigbench_geometric_shapes | 0|multiple_choice_grade|0.1003|± |0.0159|
80
+ | | |exact_str_match |0.0000|± |0.0000|
81
+ |bigbench_logical_deduction_five_objects | 0|multiple_choice_grade|0.2520|± |0.0194|
82
+ |bigbench_logical_deduction_seven_objects | 0|multiple_choice_grade|0.1871|± |0.0148|
83
+ |bigbench_logical_deduction_three_objects | 0|multiple_choice_grade|0.3833|± |0.0281|
84
+ |bigbench_movie_recommendation | 0|multiple_choice_grade|0.2500|± |0.0194|
85
+ |bigbench_navigate | 0|multiple_choice_grade|0.5000|± |0.0158|
86
+ |bigbench_reasoning_about_colored_objects | 0|multiple_choice_grade|0.4370|± |0.0111|
87
+ |bigbench_ruin_names | 0|multiple_choice_grade|0.2679|± |0.0209|
88
+ |bigbench_salient_translation_error_detection | 0|multiple_choice_grade|0.2495|± |0.0137|
89
+ |bigbench_snarks | 0|multiple_choice_grade|0.5249|± |0.0372|
90
+ |bigbench_sports_understanding | 0|multiple_choice_grade|0.5406|± |0.0159|
91
+ |bigbench_temporal_sequences | 0|multiple_choice_grade|0.2470|± |0.0136|
92
+ |bigbench_tracking_shuffled_objects_five_objects | 0|multiple_choice_grade|0.1944|± |0.0112|
93
+ |bigbench_tracking_shuffled_objects_seven_objects| 0|multiple_choice_grade|0.1509|± |0.0086|
94
+ |bigbench_tracking_shuffled_objects_three_objects| 0|multiple_choice_grade|0.3833|± |0.0281|
95
+ Average: 0.3367
96
+ ```
97
+
98
  TruthfulQA:
99
  ```
100
  hf-causal-experimental (pretrained=teknium/OpenHermes-7B,dtype=float16), limit: None, provide_description: False, num_fewshot: 0, batch_size: 8