tiedeman commited on
Commit
bb1ef83
1 Parent(s): 41d84eb

added a note about performance issues for low-resource languages

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -1093,7 +1093,7 @@ model-index:
1093
 
1094
  ## Model Details
1095
 
1096
- Neural machine translation model for translating from Multiple languages (mul) to unknown (deu+eng+nld).
1097
 
1098
  This model is part of the [OPUS-MT project](https://github.com/Helsinki-NLP/Opus-MT), an effort to make neural machine translation models widely available and accessible for many languages in the world. All models are originally trained using the amazing framework of [Marian NMT](https://marian-nmt.github.io/), an efficient NMT implementation written in pure C++. The models have been converted to pyTorch using the transformers library by huggingface. Training data is taken from [OPUS](https://opus.nlpl.eu/) and training pipelines use the procedures of [OPUS-MT-train](https://github.com/Helsinki-NLP/Opus-MT-train).
1099
  **Model Description:**
@@ -1126,6 +1126,8 @@ This model can be used for translation and text-to-text generation.
1126
 
1127
  Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)).
1128
 
 
 
1129
  ## How to Get Started With the Model
1130
 
1131
  A short example code:
 
1093
 
1094
  ## Model Details
1095
 
1096
+ Neural machine translation model for translating from Multiple languages (mul) to unknown (deu+eng+nld). Note that many of the listed languages will not be well supported by the model as the training data is very limited for the majority of the languages. Translation performance varies a lot and for a large number of language pairs it will not work at all.
1097
 
1098
  This model is part of the [OPUS-MT project](https://github.com/Helsinki-NLP/Opus-MT), an effort to make neural machine translation models widely available and accessible for many languages in the world. All models are originally trained using the amazing framework of [Marian NMT](https://marian-nmt.github.io/), an efficient NMT implementation written in pure C++. The models have been converted to pyTorch using the transformers library by huggingface. Training data is taken from [OPUS](https://opus.nlpl.eu/) and training pipelines use the procedures of [OPUS-MT-train](https://github.com/Helsinki-NLP/Opus-MT-train).
1099
  **Model Description:**
 
1126
 
1127
  Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)).
1128
 
1129
+ Also note that many of the listed languages will not be well supported by the model as the training data is very limited for the majority of the languages. Translation performance varies a lot and for a large number of language pairs it will not work at all.
1130
+
1131
  ## How to Get Started With the Model
1132
 
1133
  A short example code: