DeepLangLvcc
commited on
Commit
•
5646c50
1
Parent(s):
4e079f6
update readme
Browse files- README.md +5 -4
- README_EN.md +5 -4
README.md
CHANGED
@@ -70,12 +70,13 @@ LingoWhale-8B模型对学术研究完全开放,使用方通过邮件申请并
|
|
70 |
| **GPT-4** | 68.4 | 83.9 | 70.3 | 66.2 | 69.5 | 90.0 | 75.1 | 63.3 |
|
71 |
| **GPT-3.5 Turbo** | 51.1 | 68.5 | 54.1 | 47.1 | 52.4 | 57.8 | 61.6 | 46.1 |
|
72 |
| **LLaMA2-7B** | 28.9 | 45.7 | 31.4 | 26.0 | 12.8 | 16.2 | 39.2 | 26.5 |
|
73 |
-
| **ChatGLM2-6B
|
74 |
-
| **Baichuan2-7B-Base
|
75 |
-
| **Qwen-7B v1.1
|
76 |
| **LingoWhale-8B-base** | 63.6 | 60.2 | 62.8 | 50.3 | 32.9 | 55.0 | 47.5 | 43.8 |
|
77 |
|
78 |
-
|
|
|
79 |
|
80 |
# 生成样例
|
81 |
|
|
|
70 |
| **GPT-4** | 68.4 | 83.9 | 70.3 | 66.2 | 69.5 | 90.0 | 75.1 | 63.3 |
|
71 |
| **GPT-3.5 Turbo** | 51.1 | 68.5 | 54.1 | 47.1 | 52.4 | 57.8 | 61.6 | 46.1 |
|
72 |
| **LLaMA2-7B** | 28.9 | 45.7 | 31.4 | 26.0 | 12.8 | 16.2 | 39.2 | 26.5 |
|
73 |
+
| **ChatGLM2-6B**$\ast$ | 51.7 | 47.9 | - | - | - | 32.4 | 33.7 | - |
|
74 |
+
| **Baichuan2-7B-Base**$\ast$ | 54.0 | 54.2 | 57.1 | 47.5 | 18.3 | 24.5 | 41.6 | 42.7 |
|
75 |
+
| **Qwen-7B v1.1**$\ast$ | 63.5 | 58.2 | 62.2 | - | 29.9 | 51.7 | 45.0 | - |
|
76 |
| **LingoWhale-8B-base** | 63.6 | 60.2 | 62.8 | 50.3 | 32.9 | 55.0 | 47.5 | 43.8 |
|
77 |
|
78 |
+
|
79 |
+
$\textcolor{gray}\ast$<span style="color:gray">表示其模型结果来自于官方, 所有的结果都精确到小数点后1位。 </span>
|
80 |
|
81 |
# 生成样例
|
82 |
|
README_EN.md
CHANGED
@@ -68,12 +68,13 @@ These evaluation benchmarks provide standardized tests and metrics to assess lan
|
|
68 |
| **GPT-4** | 68.4 | 83.9 | 70.3 | 66.2 | 69.5 | 90.0 | 75.1 | 63.3 |
|
69 |
| **GPT-3.5 Turbo** | 51.1 | 68.5 | 54.1 | 47.1 | 52.4 | 57.8 | 61.6 | 46.1 |
|
70 |
| **LLaMA2-7B** | 28.9 | 45.7 | 31.4 | 26.0 | 12.8 | 16.2 | 39.2 | 26.5 |
|
71 |
-
| **ChatGLM2-6B
|
72 |
-
| **Baichuan2-7B-Base
|
73 |
-
| **Qwen-7B v1.1
|
74 |
| **LingoWhale-8B-base** | 63.6 | 60.2 | 62.8 | 50.3 | 32.9 | 55.0 | 47.5 | 43.8 |
|
75 |
|
76 |
-
|
|
|
77 |
|
78 |
# Generated Examples
|
79 |
|
|
|
68 |
| **GPT-4** | 68.4 | 83.9 | 70.3 | 66.2 | 69.5 | 90.0 | 75.1 | 63.3 |
|
69 |
| **GPT-3.5 Turbo** | 51.1 | 68.5 | 54.1 | 47.1 | 52.4 | 57.8 | 61.6 | 46.1 |
|
70 |
| **LLaMA2-7B** | 28.9 | 45.7 | 31.4 | 26.0 | 12.8 | 16.2 | 39.2 | 26.5 |
|
71 |
+
| **ChatGLM2-6B**$\ast$ | 51.7 | 47.9 | - | - | - | 32.4 | 33.7 | - |
|
72 |
+
| **Baichuan2-7B-Base**$\ast$ | 54.0 | 54.2 | 57.1 | 47.5 | 18.3 | 24.5 | 41.6 | 42.7 |
|
73 |
+
| **Qwen-7B v1.1**$\ast$ | 63.5 | 58.2 | 62.2 | - | 29.9 | 51.7 | 45.0 | - |
|
74 |
| **LingoWhale-8B-base** | 63.6 | 60.2 | 62.8 | 50.3 | 32.9 | 55.0 | 47.5 | 43.8 |
|
75 |
|
76 |
+
|
77 |
+
$\textcolor{gray}\ast$<span style="color:gray">indicates that the model results are from the official, and all the results are accurate to 1 decimal place. </span>
|
78 |
|
79 |
# Generated Examples
|
80 |
|