neo-nlp-dev
commited on
Commit
•
a970019
1
Parent(s):
932e28f
Update README.md
Browse files
README.md
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
-
---
|
2 |
-
library_name: transformers
|
3 |
-
license: cc-by-4.0
|
4 |
-
datasets:
|
5 |
-
- uonlp/CulturaX
|
6 |
-
---
|
7 |
|
8 |
# Model Card for Model ID
|
9 |
|
@@ -16,7 +16,7 @@ datasets:
|
|
16 |
- **Developed by:** DICE Research Group (https://dice-research.org/) @ Paderborn University (https://www.uni-paderborn.de/)
|
17 |
- **Model type:** GPT2 style (decoder-only) with alternating sparse Mixture-of-Experts layers
|
18 |
- **Number of Experts**: 16
|
19 |
-
- **Model Size**: 1.3 Billion (active) / 7.4 Billion (total)
|
20 |
- **Language(s) (NLP):** 160+
|
21 |
- **License:** CC BY 4.0 (https://creativecommons.org/licenses/by/4.0/)
|
22 |
- **Repository:** https://github.com/dice-group/LOLA
|
|
|
1 |
+
---
|
2 |
+
library_name: transformers
|
3 |
+
license: cc-by-4.0
|
4 |
+
datasets:
|
5 |
+
- uonlp/CulturaX
|
6 |
+
---
|
7 |
|
8 |
# Model Card for Model ID
|
9 |
|
|
|
16 |
- **Developed by:** DICE Research Group (https://dice-research.org/) @ Paderborn University (https://www.uni-paderborn.de/)
|
17 |
- **Model type:** GPT2 style (decoder-only) with alternating sparse Mixture-of-Experts layers
|
18 |
- **Number of Experts**: 16
|
19 |
+
- **Model Size**: 1.3 Billion (active*) / 7.4 Billion (total)
|
20 |
- **Language(s) (NLP):** 160+
|
21 |
- **License:** CC BY 4.0 (https://creativecommons.org/licenses/by/4.0/)
|
22 |
- **Repository:** https://github.com/dice-group/LOLA
|