Update README.md
Browse files
README.md
CHANGED
@@ -85,9 +85,22 @@ Currently, we are working on more suitable benchmarks for Spanish, French, Germa
|
|
85 |
|
86 |
<details>
|
87 |
<summary>Evaluation results</summary>
|
88 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
89 |
### German
|
90 |
-
|
91 |
| | arc_challenge_de | belebele_de | hellaswag_de | mmlu_de | truthfulqa_de | avg |
|
92 |
|:-------------------------------------|-------------------:|--------------:|---------------:|----------:|----------------:|---------:|
|
93 |
| occiglot/occiglot-7b-eu5 | 0.493584 | 0.646667 | 0.666631 | 0.483406 | 0.251269 | 0.508311 |
|
@@ -98,6 +111,7 @@ Currently, we are working on more suitable benchmarks for Spanish, French, Germa
|
|
98 |
| mistralai/Mistral-7B-v0.1 | 0.476476 | 0.738889 | 0.610589 | 0.529567 | 0.284264 | 0.527957 |
|
99 |
| mistralai/Mistral-7B-Instruct-v0.2 | 0.485885 | 0.688889 | 0.622438 | 0.501961 | 0.376904 | 0.535215 |
|
100 |
|
|
|
101 |
</details>
|
102 |
|
103 |
## Acknowledgements
|
|
|
85 |
|
86 |
<details>
|
87 |
<summary>Evaluation results</summary>
|
88 |
+
|
89 |
+
### English
|
90 |
+
|
91 |
+
| | arc_challenge | belebele | hellaswag | mmlu | truthfulqa | avg |
|
92 |
+
|:-------------------------------------|----------------:|-----------:|------------:|---------:|-------------:|---------:|
|
93 |
+
| occiglot/occiglot-7b-eu5 | 0.530717 | 0.726667 | 0.789882 | 0.531904 | 0.403678 | 0.59657 |
|
94 |
+
| occiglot/occiglot-7b-eu5-instruct | 0.558874 | 0.746667 | 0.799841 | 0.535109 | 0.449034 | 0.617905 |
|
95 |
+
| occiglot/occiglot-7b-de-en | 0.556314 | 0.791111 | 0.803824 | 0.568438 | 0.423251 | 0.628587 |
|
96 |
+
| occiglot/occiglot-7b-de-en-instruct | 0.604096 | 0.812222 | 0.80004 | 0.570574 | 0.493807 | 0.656148 |
|
97 |
+
| LeoLM/leo-mistral-hessianai-7b | 0.522184 | 0.736667 | 0.777833 | 0.538812 | 0.429248 | 0.600949 |
|
98 |
+
| mistralai/Mistral-7B-v0.1 | 0.612628 | 0.844444 | 0.834097 | 0.624555 | 0.426201 | 0.668385 |
|
99 |
+
| mistralai/Mistral-7B-Instruct-v0.2 | 0.637372 | 0.824444 | 0.846345 | 0.59201 | 0.668116 | 0.713657 |
|
100 |
+
|
101 |
+
|
102 |
### German
|
103 |
+
|
104 |
| | arc_challenge_de | belebele_de | hellaswag_de | mmlu_de | truthfulqa_de | avg |
|
105 |
|:-------------------------------------|-------------------:|--------------:|---------------:|----------:|----------------:|---------:|
|
106 |
| occiglot/occiglot-7b-eu5 | 0.493584 | 0.646667 | 0.666631 | 0.483406 | 0.251269 | 0.508311 |
|
|
|
111 |
| mistralai/Mistral-7B-v0.1 | 0.476476 | 0.738889 | 0.610589 | 0.529567 | 0.284264 | 0.527957 |
|
112 |
| mistralai/Mistral-7B-Instruct-v0.2 | 0.485885 | 0.688889 | 0.622438 | 0.501961 | 0.376904 | 0.535215 |
|
113 |
|
114 |
+
|
115 |
</details>
|
116 |
|
117 |
## Acknowledgements
|