BossRui commited on
Commit
779efda
1 Parent(s): 67237cd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -27
README.md CHANGED
@@ -31,7 +31,7 @@ language:
31
  - sq
32
  - da
33
  - sa
34
- - 'no'
35
  - gn
36
  - sr
37
  - sk
@@ -59,22 +59,32 @@ tags:
59
  # Democratizing Medical LLMs For Much More Languages
60
 
61
  Covering 12 Major Languages including English, Chinese, French, Hindi, Spanish, Arabic, Russian, Japanese, Korean, German, Italian, Portuguese and 38 Minor Languages So far.
 
62
 
63
 
64
 
65
  <p align="center">
66
- 📃 <a href="https://arxiv.org/abs/2410.10626" target="_blank">Paper</a> • 🌐 <a href="" target="_blank">Demo</a> • 🤗 <a href="https://huggingface.co/datasets/FreedomIntelligence/ApolloMoEDataset" target="_blank">ApolloMoEDataset</a> • 🤗 <a href="https://huggingface.co/datasets/FreedomIntelligence/ApolloMoEBench" target="_blank">ApolloMoEBench</a> • 🤗 <a href="https://huggingface.co/collections/FreedomIntelligence/apollomoe-and-apollo2-670ddebe3bb1ba1aebabbf2c" target="_blank">Models</a> • 🌐 <a href="https://github.com/FreedomIntelligence/Apollo" target="_blank">Apollo</a> • 🌐 <a href="https://github.com/FreedomIntelligence/ApolloMoE" target="_blank">ApolloMoE</a>
67
  </p>
68
 
69
-
70
  ![Apollo](assets/apollo_medium_final.png)
71
 
72
-
73
  ## 🌈 Update
74
 
75
  * **[2024.10.15]** ApolloMoE repo is published!🎉
76
 
77
 
 
 
 
 
 
 
 
 
 
 
 
78
  ## Architecture
79
 
80
  <details>
@@ -105,12 +115,7 @@ Covering 12 Major Languages including English, Chinese, French, Hindi, Spanish,
105
  ![ApolloMoE](assets/post_moe_results.png)
106
 
107
  </details>
108
-
109
-
110
-
111
 
112
-
113
-
114
 
115
  ## Usage Format
116
  #### Apollo2
@@ -120,7 +125,7 @@ Covering 12 Major Languages including English, Chinese, French, Hindi, Spanish,
120
 
121
  #### Apollo-MoE
122
  - 0.5B, 1.5B, 7B: User:{query}\nAssistant:{response}<|endoftext|>
123
-
124
  ## Dataset & Evaluation
125
 
126
  - Dataset
@@ -176,9 +181,6 @@ Covering 12 Major Languages including English, Chinese, French, Hindi, Spanish,
176
  - RU: [RuMedBench](https://github.com/sb-ai-lab/MedBench)
177
 
178
 
179
-
180
-
181
-
182
  </details>
183
 
184
 
@@ -186,17 +188,17 @@ Covering 12 Major Languages including English, Chinese, French, Hindi, Spanish,
186
  <details><summary>Click to expand</summary>
187
 
188
 
189
- We take Gemma-2b as example
190
  1. Download Dataset for project:
191
 
192
  ```
193
- bash 0.download_data.sh
194
  ```
195
 
196
- 2. Prepare test and dev for specific model:
197
 
198
 
199
- - Create test data for with special token, you can use ./util/check.ipynb to check models' special tokens
200
 
201
  ```
202
  bash 1.data_process_test&dev.sh
@@ -214,13 +216,11 @@ Covering 12 Major Languages including English, Chinese, French, Hindi, Spanish,
214
  4. Train the model
215
 
216
 
217
- - If you want to train in Multi Nodes please refer to ./scripts/multi_node_train_*.sh
218
-
219
-
220
 
221
 
222
  ```
223
- bash 3.single_node_train_gemma.sh
224
  ```
225
 
226
 
@@ -230,12 +230,6 @@ Covering 12 Major Languages including English, Chinese, French, Hindi, Spanish,
230
  bash 4.eval.sh
231
  ```
232
 
233
- 6. Evaluate your model: Play with your ckpts in bash
234
-
235
- ```
236
- python ./src/evaluate/cli_demo.py --model_name='./ckpts/your/path/tfmr'
237
- ```
238
-
239
  </details>
240
 
241
 
 
31
  - sq
32
  - da
33
  - sa
34
+ - no
35
  - gn
36
  - sr
37
  - sk
 
59
  # Democratizing Medical LLMs For Much More Languages
60
 
61
  Covering 12 Major Languages including English, Chinese, French, Hindi, Spanish, Arabic, Russian, Japanese, Korean, German, Italian, Portuguese and 38 Minor Languages So far.
62
+ <center>
63
 
64
 
65
 
66
  <p align="center">
67
+ 📃 <a href="https://arxiv.org/abs/2410.10626" target="_blank">Paper</a> • 🌐 <a href="" target="_blank">Demo</a> • 🤗 <a href="https://huggingface.co/datasets/FreedomIntelligence/ApolloMoEDataset" target="_blank">ApolloMoEDataset</a> • 🤗 <a href="https://huggingface.co/datasets/FreedomIntelligence/ApolloMoEBench" target="_blank">ApolloMoEBench</a> • 🤗 <a href="https://huggingface.co/collections/FreedomIntelligence/apollomoe-and-apollo2-670ddebe3bb1ba1aebabbf2c" target="_blank">Models</a> • 🌐 <a href="https://github.com/FreedomIntelligence/Apollo" target="_blank">Apollo</a>
68
  </p>
69
 
 
70
  ![Apollo](assets/apollo_medium_final.png)
71
 
 
72
  ## 🌈 Update
73
 
74
  * **[2024.10.15]** ApolloMoE repo is published!🎉
75
 
76
 
77
+ ## Languages Coverage
78
+ 12 Major Languages and 38 Minor Languages
79
+
80
+ <details>
81
+ <summary>Click to view the Languages Coverage</summary>
82
+
83
+ ![ApolloMoE](assets/languages.png)
84
+
85
+ </details>
86
+
87
+
88
  ## Architecture
89
 
90
  <details>
 
115
  ![ApolloMoE](assets/post_moe_results.png)
116
 
117
  </details>
 
 
 
118
 
 
 
119
 
120
  ## Usage Format
121
  #### Apollo2
 
125
 
126
  #### Apollo-MoE
127
  - 0.5B, 1.5B, 7B: User:{query}\nAssistant:{response}<|endoftext|>
128
+
129
  ## Dataset & Evaluation
130
 
131
  - Dataset
 
181
  - RU: [RuMedBench](https://github.com/sb-ai-lab/MedBench)
182
 
183
 
 
 
 
184
  </details>
185
 
186
 
 
188
  <details><summary>Click to expand</summary>
189
 
190
 
191
+ We take Apollo2-7B or Apollo-MoE-0.5B as example
192
  1. Download Dataset for project:
193
 
194
  ```
195
+ bash 0.download_data.sh 
196
  ```
197
 
198
+ 2. Prepare test and dev data for specific model:
199
 
200
 
201
+ - Create test data for with special token
202
 
203
  ```
204
  bash 1.data_process_test&dev.sh
 
216
  4. Train the model
217
 
218
 
219
+ - If you want to train in Multi Nodes please refer to ./src/sft/training_config/zero_multi.yaml
 
 
220
 
221
 
222
  ```
223
+ bash 3.single_node_train.sh
224
  ```
225
 
226
 
 
230
  bash 4.eval.sh
231
  ```
232
 
 
 
 
 
 
 
233
  </details>
234
 
235