Suparious commited on
Commit
3918370
1 Parent(s): 38c12c9

add eval results

Browse files
Files changed (1) hide show
  1. README.md +103 -0
README.md CHANGED
@@ -18,6 +18,109 @@ tags:
18
  - endpoints_compatible
19
  - text-generation-inference
20
  - chatml
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21
  base_model:
22
  - Locutusque/Hercules-2.5-Mistral-7B
23
  - openchat/openchat-3.5-0106
 
18
  - endpoints_compatible
19
  - text-generation-inference
20
  - chatml
21
+ model-index:
22
+ - name: ChatHercules-2.5-Mistral-7B
23
+ results:
24
+ - task:
25
+ type: text-generation
26
+ name: Text Generation
27
+ dataset:
28
+ name: AI2 Reasoning Challenge (25-Shot)
29
+ type: ai2_arc
30
+ config: ARC-Challenge
31
+ split: test
32
+ args:
33
+ num_few_shot: 25
34
+ metrics:
35
+ - type: acc_norm
36
+ value: 65.1
37
+ name: normalized accuracy
38
+ source:
39
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=hydra-project/ChatHercules-2.5-Mistral-7B
40
+ name: Open LLM Leaderboard
41
+ - task:
42
+ type: text-generation
43
+ name: Text Generation
44
+ dataset:
45
+ name: HellaSwag (10-Shot)
46
+ type: hellaswag
47
+ split: validation
48
+ args:
49
+ num_few_shot: 10
50
+ metrics:
51
+ - type: acc_norm
52
+ value: 84.61
53
+ name: normalized accuracy
54
+ source:
55
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=hydra-project/ChatHercules-2.5-Mistral-7B
56
+ name: Open LLM Leaderboard
57
+ - task:
58
+ type: text-generation
59
+ name: Text Generation
60
+ dataset:
61
+ name: MMLU (5-Shot)
62
+ type: cais/mmlu
63
+ config: all
64
+ split: test
65
+ args:
66
+ num_few_shot: 5
67
+ metrics:
68
+ - type: acc
69
+ value: 65.35
70
+ name: accuracy
71
+ source:
72
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=hydra-project/ChatHercules-2.5-Mistral-7B
73
+ name: Open LLM Leaderboard
74
+ - task:
75
+ type: text-generation
76
+ name: Text Generation
77
+ dataset:
78
+ name: TruthfulQA (0-shot)
79
+ type: truthful_qa
80
+ config: multiple_choice
81
+ split: validation
82
+ args:
83
+ num_few_shot: 0
84
+ metrics:
85
+ - type: mc2
86
+ value: 47.52
87
+ source:
88
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=hydra-project/ChatHercules-2.5-Mistral-7B
89
+ name: Open LLM Leaderboard
90
+ - task:
91
+ type: text-generation
92
+ name: Text Generation
93
+ dataset:
94
+ name: Winogrande (5-shot)
95
+ type: winogrande
96
+ config: winogrande_xl
97
+ split: validation
98
+ args:
99
+ num_few_shot: 5
100
+ metrics:
101
+ - type: acc
102
+ value: 81.85
103
+ name: accuracy
104
+ source:
105
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=hydra-project/ChatHercules-2.5-Mistral-7B
106
+ name: Open LLM Leaderboard
107
+ - task:
108
+ type: text-generation
109
+ name: Text Generation
110
+ dataset:
111
+ name: GSM8k (5-shot)
112
+ type: gsm8k
113
+ config: main
114
+ split: test
115
+ args:
116
+ num_few_shot: 5
117
+ metrics:
118
+ - type: acc
119
+ value: 64.97
120
+ name: accuracy
121
+ source:
122
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=hydra-project/ChatHercules-2.5-Mistral-7B
123
+ name: Open LLM Leaderboard
124
  base_model:
125
  - Locutusque/Hercules-2.5-Mistral-7B
126
  - openchat/openchat-3.5-0106