szymonrucinski
commited on
Commit
•
06bc2b9
1
Parent(s):
f15815f
Update README.md
Browse files
README.md
CHANGED
@@ -56,5 +56,6 @@ The model was trained using AdamW optimizer, with specific hyperparameters aimed
|
|
56 |
## Conclusions
|
57 |
The `Curie-7B-v1` model, through LAPT, matches foundational models on eight downstream tasks with significantly less data. Its versatility in generating Polish text and the ability to be transformed into classifiers, regressors, and AI assistants highlights the method's effectiveness. This open-source Polish LLM provides a foundation for developing efficient business solutions.
|
58 |
|
59 |
-
##
|
60 |
-
|
|
|
|
56 |
## Conclusions
|
57 |
The `Curie-7B-v1` model, through LAPT, matches foundational models on eight downstream tasks with significantly less data. Its versatility in generating Polish text and the ability to be transformed into classifiers, regressors, and AI assistants highlights the method's effectiveness. This open-source Polish LLM provides a foundation for developing efficient business solutions.
|
58 |
|
59 |
+
## Research Paper
|
60 |
+
Work and details regarding this model are described in the reserach paper [Efficient Language Adaptive Pre-training: Extending State-of-the-Art Large Language Models for Polish](https://arxiv.org/abs/2402.09759) by Szymon Ruciński.
|
61 |
+
|