added a note about performance issues for low-resource languages
Browse files
README.md
CHANGED
@@ -1093,7 +1093,7 @@ model-index:
|
|
1093 |
|
1094 |
## Model Details
|
1095 |
|
1096 |
-
Neural machine translation model for translating from Multiple languages (mul) to unknown (deu+eng+nld).
|
1097 |
|
1098 |
This model is part of the [OPUS-MT project](https://github.com/Helsinki-NLP/Opus-MT), an effort to make neural machine translation models widely available and accessible for many languages in the world. All models are originally trained using the amazing framework of [Marian NMT](https://marian-nmt.github.io/), an efficient NMT implementation written in pure C++. The models have been converted to pyTorch using the transformers library by huggingface. Training data is taken from [OPUS](https://opus.nlpl.eu/) and training pipelines use the procedures of [OPUS-MT-train](https://github.com/Helsinki-NLP/Opus-MT-train).
|
1099 |
**Model Description:**
|
@@ -1126,6 +1126,8 @@ This model can be used for translation and text-to-text generation.
|
|
1126 |
|
1127 |
Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)).
|
1128 |
|
|
|
|
|
1129 |
## How to Get Started With the Model
|
1130 |
|
1131 |
A short example code:
|
|
|
1093 |
|
1094 |
## Model Details
|
1095 |
|
1096 |
+
Neural machine translation model for translating from Multiple languages (mul) to unknown (deu+eng+nld). Note that many of the listed languages will not be well supported by the model as the training data is very limited for the majority of the languages. Translation performance varies a lot and for a large number of language pairs it will not work at all.
|
1097 |
|
1098 |
This model is part of the [OPUS-MT project](https://github.com/Helsinki-NLP/Opus-MT), an effort to make neural machine translation models widely available and accessible for many languages in the world. All models are originally trained using the amazing framework of [Marian NMT](https://marian-nmt.github.io/), an efficient NMT implementation written in pure C++. The models have been converted to pyTorch using the transformers library by huggingface. Training data is taken from [OPUS](https://opus.nlpl.eu/) and training pipelines use the procedures of [OPUS-MT-train](https://github.com/Helsinki-NLP/Opus-MT-train).
|
1099 |
**Model Description:**
|
|
|
1126 |
|
1127 |
Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)).
|
1128 |
|
1129 |
+
Also note that many of the listed languages will not be well supported by the model as the training data is very limited for the majority of the languages. Translation performance varies a lot and for a large number of language pairs it will not work at all.
|
1130 |
+
|
1131 |
## How to Get Started With the Model
|
1132 |
|
1133 |
A short example code:
|