Initial commit
Browse files- .gitattributes +1 -0
- README.md +2186 -0
- benchmark_results.txt +263 -0
- benchmark_translations.zip +0 -0
- config.json +41 -0
- generation_config.json +16 -0
- model.safetensors +3 -0
- pytorch_model.bin +3 -0
- source.spm +3 -0
- special_tokens_map.json +1 -0
- target.spm +3 -0
- tokenizer_config.json +1 -0
- vocab.json +0 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
*.spm filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,2186 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
library_name: transformers
|
3 |
+
language:
|
4 |
+
- anp
|
5 |
+
- as
|
6 |
+
- awa
|
7 |
+
- bal
|
8 |
+
- bho
|
9 |
+
- bn
|
10 |
+
- bpy
|
11 |
+
- de
|
12 |
+
- diq
|
13 |
+
- dv
|
14 |
+
- en
|
15 |
+
- es
|
16 |
+
- fa
|
17 |
+
- fr
|
18 |
+
- gbm
|
19 |
+
- glk
|
20 |
+
- gu
|
21 |
+
- hi
|
22 |
+
- hif
|
23 |
+
- hne
|
24 |
+
- hns
|
25 |
+
- jdt
|
26 |
+
- kok
|
27 |
+
- ks
|
28 |
+
- ku
|
29 |
+
- lah
|
30 |
+
- lrc
|
31 |
+
- mag
|
32 |
+
- mai
|
33 |
+
- mr
|
34 |
+
- mzn
|
35 |
+
- ne
|
36 |
+
- or
|
37 |
+
- os
|
38 |
+
- pa
|
39 |
+
- pal
|
40 |
+
- pi
|
41 |
+
- ps
|
42 |
+
- pt
|
43 |
+
- rhg
|
44 |
+
- rmy
|
45 |
+
- rom
|
46 |
+
- sa
|
47 |
+
- sd
|
48 |
+
- si
|
49 |
+
- skr
|
50 |
+
- syl
|
51 |
+
- tg
|
52 |
+
- tly
|
53 |
+
- ur
|
54 |
+
- zza
|
55 |
+
|
56 |
+
tags:
|
57 |
+
- translation
|
58 |
+
- opus-mt-tc-bible
|
59 |
+
|
60 |
+
license: apache-2.0
|
61 |
+
model-index:
|
62 |
+
- name: opus-mt-tc-bible-big-deu_eng_fra_por_spa-iir
|
63 |
+
results:
|
64 |
+
- task:
|
65 |
+
name: Translation deu-ben
|
66 |
+
type: translation
|
67 |
+
args: deu-ben
|
68 |
+
dataset:
|
69 |
+
name: flores200-devtest
|
70 |
+
type: flores200-devtest
|
71 |
+
args: deu-ben
|
72 |
+
metrics:
|
73 |
+
- name: BLEU
|
74 |
+
type: bleu
|
75 |
+
value: 10.6
|
76 |
+
- name: chr-F
|
77 |
+
type: chrf
|
78 |
+
value: 0.44005
|
79 |
+
- task:
|
80 |
+
name: Translation deu-guj
|
81 |
+
type: translation
|
82 |
+
args: deu-guj
|
83 |
+
dataset:
|
84 |
+
name: flores200-devtest
|
85 |
+
type: flores200-devtest
|
86 |
+
args: deu-guj
|
87 |
+
metrics:
|
88 |
+
- name: BLEU
|
89 |
+
type: bleu
|
90 |
+
value: 11.1
|
91 |
+
- name: chr-F
|
92 |
+
type: chrf
|
93 |
+
value: 0.39522
|
94 |
+
- task:
|
95 |
+
name: Translation deu-hin
|
96 |
+
type: translation
|
97 |
+
args: deu-hin
|
98 |
+
dataset:
|
99 |
+
name: flores200-devtest
|
100 |
+
type: flores200-devtest
|
101 |
+
args: deu-hin
|
102 |
+
metrics:
|
103 |
+
- name: BLEU
|
104 |
+
type: bleu
|
105 |
+
value: 22.3
|
106 |
+
- name: chr-F
|
107 |
+
type: chrf
|
108 |
+
value: 0.48448
|
109 |
+
- task:
|
110 |
+
name: Translation deu-hne
|
111 |
+
type: translation
|
112 |
+
args: deu-hne
|
113 |
+
dataset:
|
114 |
+
name: flores200-devtest
|
115 |
+
type: flores200-devtest
|
116 |
+
args: deu-hne
|
117 |
+
metrics:
|
118 |
+
- name: BLEU
|
119 |
+
type: bleu
|
120 |
+
value: 13.8
|
121 |
+
- name: chr-F
|
122 |
+
type: chrf
|
123 |
+
value: 0.42659
|
124 |
+
- task:
|
125 |
+
name: Translation deu-mag
|
126 |
+
type: translation
|
127 |
+
args: deu-mag
|
128 |
+
dataset:
|
129 |
+
name: flores200-devtest
|
130 |
+
type: flores200-devtest
|
131 |
+
args: deu-mag
|
132 |
+
metrics:
|
133 |
+
- name: BLEU
|
134 |
+
type: bleu
|
135 |
+
value: 14.0
|
136 |
+
- name: chr-F
|
137 |
+
type: chrf
|
138 |
+
value: 0.42477
|
139 |
+
- task:
|
140 |
+
name: Translation deu-pan
|
141 |
+
type: translation
|
142 |
+
args: deu-pan
|
143 |
+
dataset:
|
144 |
+
name: flores200-devtest
|
145 |
+
type: flores200-devtest
|
146 |
+
args: deu-pan
|
147 |
+
metrics:
|
148 |
+
- name: BLEU
|
149 |
+
type: bleu
|
150 |
+
value: 11.3
|
151 |
+
- name: chr-F
|
152 |
+
type: chrf
|
153 |
+
value: 0.37308
|
154 |
+
- task:
|
155 |
+
name: Translation deu-pes
|
156 |
+
type: translation
|
157 |
+
args: deu-pes
|
158 |
+
dataset:
|
159 |
+
name: flores200-devtest
|
160 |
+
type: flores200-devtest
|
161 |
+
args: deu-pes
|
162 |
+
metrics:
|
163 |
+
- name: BLEU
|
164 |
+
type: bleu
|
165 |
+
value: 14.9
|
166 |
+
- name: chr-F
|
167 |
+
type: chrf
|
168 |
+
value: 0.42726
|
169 |
+
- task:
|
170 |
+
name: Translation deu-prs
|
171 |
+
type: translation
|
172 |
+
args: deu-prs
|
173 |
+
dataset:
|
174 |
+
name: flores200-devtest
|
175 |
+
type: flores200-devtest
|
176 |
+
args: deu-prs
|
177 |
+
metrics:
|
178 |
+
- name: BLEU
|
179 |
+
type: bleu
|
180 |
+
value: 12.9
|
181 |
+
- name: chr-F
|
182 |
+
type: chrf
|
183 |
+
value: 0.38830
|
184 |
+
- task:
|
185 |
+
name: Translation deu-tgk
|
186 |
+
type: translation
|
187 |
+
args: deu-tgk
|
188 |
+
dataset:
|
189 |
+
name: flores200-devtest
|
190 |
+
type: flores200-devtest
|
191 |
+
args: deu-tgk
|
192 |
+
metrics:
|
193 |
+
- name: BLEU
|
194 |
+
type: bleu
|
195 |
+
value: 12.9
|
196 |
+
- name: chr-F
|
197 |
+
type: chrf
|
198 |
+
value: 0.40932
|
199 |
+
- task:
|
200 |
+
name: Translation deu-urd
|
201 |
+
type: translation
|
202 |
+
args: deu-urd
|
203 |
+
dataset:
|
204 |
+
name: flores200-devtest
|
205 |
+
type: flores200-devtest
|
206 |
+
args: deu-urd
|
207 |
+
metrics:
|
208 |
+
- name: BLEU
|
209 |
+
type: bleu
|
210 |
+
value: 14.4
|
211 |
+
- name: chr-F
|
212 |
+
type: chrf
|
213 |
+
value: 0.41250
|
214 |
+
- task:
|
215 |
+
name: Translation eng-ben
|
216 |
+
type: translation
|
217 |
+
args: eng-ben
|
218 |
+
dataset:
|
219 |
+
name: flores200-devtest
|
220 |
+
type: flores200-devtest
|
221 |
+
args: eng-ben
|
222 |
+
metrics:
|
223 |
+
- name: BLEU
|
224 |
+
type: bleu
|
225 |
+
value: 17.1
|
226 |
+
- name: chr-F
|
227 |
+
type: chrf
|
228 |
+
value: 0.51361
|
229 |
+
- task:
|
230 |
+
name: Translation eng-bho
|
231 |
+
type: translation
|
232 |
+
args: eng-bho
|
233 |
+
dataset:
|
234 |
+
name: flores200-devtest
|
235 |
+
type: flores200-devtest
|
236 |
+
args: eng-bho
|
237 |
+
metrics:
|
238 |
+
- name: BLEU
|
239 |
+
type: bleu
|
240 |
+
value: 12.1
|
241 |
+
- name: chr-F
|
242 |
+
type: chrf
|
243 |
+
value: 0.38188
|
244 |
+
- task:
|
245 |
+
name: Translation eng-guj
|
246 |
+
type: translation
|
247 |
+
args: eng-guj
|
248 |
+
dataset:
|
249 |
+
name: flores200-devtest
|
250 |
+
type: flores200-devtest
|
251 |
+
args: eng-guj
|
252 |
+
metrics:
|
253 |
+
- name: BLEU
|
254 |
+
type: bleu
|
255 |
+
value: 22.4
|
256 |
+
- name: chr-F
|
257 |
+
type: chrf
|
258 |
+
value: 0.54231
|
259 |
+
- task:
|
260 |
+
name: Translation eng-hin
|
261 |
+
type: translation
|
262 |
+
args: eng-hin
|
263 |
+
dataset:
|
264 |
+
name: flores200-devtest
|
265 |
+
type: flores200-devtest
|
266 |
+
args: eng-hin
|
267 |
+
metrics:
|
268 |
+
- name: BLEU
|
269 |
+
type: bleu
|
270 |
+
value: 33.7
|
271 |
+
- name: chr-F
|
272 |
+
type: chrf
|
273 |
+
value: 0.58371
|
274 |
+
- task:
|
275 |
+
name: Translation eng-hne
|
276 |
+
type: translation
|
277 |
+
args: eng-hne
|
278 |
+
dataset:
|
279 |
+
name: flores200-devtest
|
280 |
+
type: flores200-devtest
|
281 |
+
args: eng-hne
|
282 |
+
metrics:
|
283 |
+
- name: BLEU
|
284 |
+
type: bleu
|
285 |
+
value: 19.9
|
286 |
+
- name: chr-F
|
287 |
+
type: chrf
|
288 |
+
value: 0.47591
|
289 |
+
- task:
|
290 |
+
name: Translation eng-mag
|
291 |
+
type: translation
|
292 |
+
args: eng-mag
|
293 |
+
dataset:
|
294 |
+
name: flores200-devtest
|
295 |
+
type: flores200-devtest
|
296 |
+
args: eng-mag
|
297 |
+
metrics:
|
298 |
+
- name: BLEU
|
299 |
+
type: bleu
|
300 |
+
value: 22.2
|
301 |
+
- name: chr-F
|
302 |
+
type: chrf
|
303 |
+
value: 0.51070
|
304 |
+
- task:
|
305 |
+
name: Translation eng-mai
|
306 |
+
type: translation
|
307 |
+
args: eng-mai
|
308 |
+
dataset:
|
309 |
+
name: flores200-devtest
|
310 |
+
type: flores200-devtest
|
311 |
+
args: eng-mai
|
312 |
+
metrics:
|
313 |
+
- name: BLEU
|
314 |
+
type: bleu
|
315 |
+
value: 10.0
|
316 |
+
- name: chr-F
|
317 |
+
type: chrf
|
318 |
+
value: 0.39249
|
319 |
+
- task:
|
320 |
+
name: Translation eng-mar
|
321 |
+
type: translation
|
322 |
+
args: eng-mar
|
323 |
+
dataset:
|
324 |
+
name: flores200-devtest
|
325 |
+
type: flores200-devtest
|
326 |
+
args: eng-mar
|
327 |
+
metrics:
|
328 |
+
- name: BLEU
|
329 |
+
type: bleu
|
330 |
+
value: 14.8
|
331 |
+
- name: chr-F
|
332 |
+
type: chrf
|
333 |
+
value: 0.48733
|
334 |
+
- task:
|
335 |
+
name: Translation eng-pan
|
336 |
+
type: translation
|
337 |
+
args: eng-pan
|
338 |
+
dataset:
|
339 |
+
name: flores200-devtest
|
340 |
+
type: flores200-devtest
|
341 |
+
args: eng-pan
|
342 |
+
metrics:
|
343 |
+
- name: BLEU
|
344 |
+
type: bleu
|
345 |
+
value: 18.1
|
346 |
+
- name: chr-F
|
347 |
+
type: chrf
|
348 |
+
value: 0.45015
|
349 |
+
- task:
|
350 |
+
name: Translation eng-pes
|
351 |
+
type: translation
|
352 |
+
args: eng-pes
|
353 |
+
dataset:
|
354 |
+
name: flores200-devtest
|
355 |
+
type: flores200-devtest
|
356 |
+
args: eng-pes
|
357 |
+
metrics:
|
358 |
+
- name: BLEU
|
359 |
+
type: bleu
|
360 |
+
value: 21.1
|
361 |
+
- name: chr-F
|
362 |
+
type: chrf
|
363 |
+
value: 0.48588
|
364 |
+
- task:
|
365 |
+
name: Translation eng-prs
|
366 |
+
type: translation
|
367 |
+
args: eng-prs
|
368 |
+
dataset:
|
369 |
+
name: flores200-devtest
|
370 |
+
type: flores200-devtest
|
371 |
+
args: eng-prs
|
372 |
+
metrics:
|
373 |
+
- name: BLEU
|
374 |
+
type: bleu
|
375 |
+
value: 24.5
|
376 |
+
- name: chr-F
|
377 |
+
type: chrf
|
378 |
+
value: 0.51879
|
379 |
+
- task:
|
380 |
+
name: Translation eng-sin
|
381 |
+
type: translation
|
382 |
+
args: eng-sin
|
383 |
+
dataset:
|
384 |
+
name: flores200-devtest
|
385 |
+
type: flores200-devtest
|
386 |
+
args: eng-sin
|
387 |
+
metrics:
|
388 |
+
- name: BLEU
|
389 |
+
type: bleu
|
390 |
+
value: 10.6
|
391 |
+
- name: chr-F
|
392 |
+
type: chrf
|
393 |
+
value: 0.43823
|
394 |
+
- task:
|
395 |
+
name: Translation eng-tgk
|
396 |
+
type: translation
|
397 |
+
args: eng-tgk
|
398 |
+
dataset:
|
399 |
+
name: flores200-devtest
|
400 |
+
type: flores200-devtest
|
401 |
+
args: eng-tgk
|
402 |
+
metrics:
|
403 |
+
- name: BLEU
|
404 |
+
type: bleu
|
405 |
+
value: 17.8
|
406 |
+
- name: chr-F
|
407 |
+
type: chrf
|
408 |
+
value: 0.47323
|
409 |
+
- task:
|
410 |
+
name: Translation eng-urd
|
411 |
+
type: translation
|
412 |
+
args: eng-urd
|
413 |
+
dataset:
|
414 |
+
name: flores200-devtest
|
415 |
+
type: flores200-devtest
|
416 |
+
args: eng-urd
|
417 |
+
metrics:
|
418 |
+
- name: BLEU
|
419 |
+
type: bleu
|
420 |
+
value: 20.4
|
421 |
+
- name: chr-F
|
422 |
+
type: chrf
|
423 |
+
value: 0.48212
|
424 |
+
- task:
|
425 |
+
name: Translation fra-ben
|
426 |
+
type: translation
|
427 |
+
args: fra-ben
|
428 |
+
dataset:
|
429 |
+
name: flores200-devtest
|
430 |
+
type: flores200-devtest
|
431 |
+
args: fra-ben
|
432 |
+
metrics:
|
433 |
+
- name: BLEU
|
434 |
+
type: bleu
|
435 |
+
value: 11.0
|
436 |
+
- name: chr-F
|
437 |
+
type: chrf
|
438 |
+
value: 0.44029
|
439 |
+
- task:
|
440 |
+
name: Translation fra-guj
|
441 |
+
type: translation
|
442 |
+
args: fra-guj
|
443 |
+
dataset:
|
444 |
+
name: flores200-devtest
|
445 |
+
type: flores200-devtest
|
446 |
+
args: fra-guj
|
447 |
+
metrics:
|
448 |
+
- name: BLEU
|
449 |
+
type: bleu
|
450 |
+
value: 11.0
|
451 |
+
- name: chr-F
|
452 |
+
type: chrf
|
453 |
+
value: 0.38634
|
454 |
+
- task:
|
455 |
+
name: Translation fra-hin
|
456 |
+
type: translation
|
457 |
+
args: fra-hin
|
458 |
+
dataset:
|
459 |
+
name: flores200-devtest
|
460 |
+
type: flores200-devtest
|
461 |
+
args: fra-hin
|
462 |
+
metrics:
|
463 |
+
- name: BLEU
|
464 |
+
type: bleu
|
465 |
+
value: 22.6
|
466 |
+
- name: chr-F
|
467 |
+
type: chrf
|
468 |
+
value: 0.48406
|
469 |
+
- task:
|
470 |
+
name: Translation fra-hne
|
471 |
+
type: translation
|
472 |
+
args: fra-hne
|
473 |
+
dataset:
|
474 |
+
name: flores200-devtest
|
475 |
+
type: flores200-devtest
|
476 |
+
args: fra-hne
|
477 |
+
metrics:
|
478 |
+
- name: BLEU
|
479 |
+
type: bleu
|
480 |
+
value: 13.9
|
481 |
+
- name: chr-F
|
482 |
+
type: chrf
|
483 |
+
value: 0.42353
|
484 |
+
- task:
|
485 |
+
name: Translation fra-mag
|
486 |
+
type: translation
|
487 |
+
args: fra-mag
|
488 |
+
dataset:
|
489 |
+
name: flores200-devtest
|
490 |
+
type: flores200-devtest
|
491 |
+
args: fra-mag
|
492 |
+
metrics:
|
493 |
+
- name: BLEU
|
494 |
+
type: bleu
|
495 |
+
value: 14.3
|
496 |
+
- name: chr-F
|
497 |
+
type: chrf
|
498 |
+
value: 0.42678
|
499 |
+
- task:
|
500 |
+
name: Translation fra-pan
|
501 |
+
type: translation
|
502 |
+
args: fra-pan
|
503 |
+
dataset:
|
504 |
+
name: flores200-devtest
|
505 |
+
type: flores200-devtest
|
506 |
+
args: fra-pan
|
507 |
+
metrics:
|
508 |
+
- name: BLEU
|
509 |
+
type: bleu
|
510 |
+
value: 10.6
|
511 |
+
- name: chr-F
|
512 |
+
type: chrf
|
513 |
+
value: 0.36980
|
514 |
+
- task:
|
515 |
+
name: Translation fra-pes
|
516 |
+
type: translation
|
517 |
+
args: fra-pes
|
518 |
+
dataset:
|
519 |
+
name: flores200-devtest
|
520 |
+
type: flores200-devtest
|
521 |
+
args: fra-pes
|
522 |
+
metrics:
|
523 |
+
- name: BLEU
|
524 |
+
type: bleu
|
525 |
+
value: 15.5
|
526 |
+
- name: chr-F
|
527 |
+
type: chrf
|
528 |
+
value: 0.43526
|
529 |
+
- task:
|
530 |
+
name: Translation fra-prs
|
531 |
+
type: translation
|
532 |
+
args: fra-prs
|
533 |
+
dataset:
|
534 |
+
name: flores200-devtest
|
535 |
+
type: flores200-devtest
|
536 |
+
args: fra-prs
|
537 |
+
metrics:
|
538 |
+
- name: BLEU
|
539 |
+
type: bleu
|
540 |
+
value: 12.8
|
541 |
+
- name: chr-F
|
542 |
+
type: chrf
|
543 |
+
value: 0.37985
|
544 |
+
- task:
|
545 |
+
name: Translation fra-tgk
|
546 |
+
type: translation
|
547 |
+
args: fra-tgk
|
548 |
+
dataset:
|
549 |
+
name: flores200-devtest
|
550 |
+
type: flores200-devtest
|
551 |
+
args: fra-tgk
|
552 |
+
metrics:
|
553 |
+
- name: BLEU
|
554 |
+
type: bleu
|
555 |
+
value: 13.7
|
556 |
+
- name: chr-F
|
557 |
+
type: chrf
|
558 |
+
value: 0.42982
|
559 |
+
- task:
|
560 |
+
name: Translation fra-urd
|
561 |
+
type: translation
|
562 |
+
args: fra-urd
|
563 |
+
dataset:
|
564 |
+
name: flores200-devtest
|
565 |
+
type: flores200-devtest
|
566 |
+
args: fra-urd
|
567 |
+
metrics:
|
568 |
+
- name: BLEU
|
569 |
+
type: bleu
|
570 |
+
value: 14.2
|
571 |
+
- name: chr-F
|
572 |
+
type: chrf
|
573 |
+
value: 0.41438
|
574 |
+
- task:
|
575 |
+
name: Translation por-ben
|
576 |
+
type: translation
|
577 |
+
args: por-ben
|
578 |
+
dataset:
|
579 |
+
name: flores200-devtest
|
580 |
+
type: flores200-devtest
|
581 |
+
args: por-ben
|
582 |
+
metrics:
|
583 |
+
- name: BLEU
|
584 |
+
type: bleu
|
585 |
+
value: 10.4
|
586 |
+
- name: chr-F
|
587 |
+
type: chrf
|
588 |
+
value: 0.43390
|
589 |
+
- task:
|
590 |
+
name: Translation por-guj
|
591 |
+
type: translation
|
592 |
+
args: por-guj
|
593 |
+
dataset:
|
594 |
+
name: flores200-devtest
|
595 |
+
type: flores200-devtest
|
596 |
+
args: por-guj
|
597 |
+
metrics:
|
598 |
+
- name: BLEU
|
599 |
+
type: bleu
|
600 |
+
value: 10.6
|
601 |
+
- name: chr-F
|
602 |
+
type: chrf
|
603 |
+
value: 0.37374
|
604 |
+
- task:
|
605 |
+
name: Translation por-hin
|
606 |
+
type: translation
|
607 |
+
args: por-hin
|
608 |
+
dataset:
|
609 |
+
name: flores200-devtest
|
610 |
+
type: flores200-devtest
|
611 |
+
args: por-hin
|
612 |
+
metrics:
|
613 |
+
- name: BLEU
|
614 |
+
type: bleu
|
615 |
+
value: 23.6
|
616 |
+
- name: chr-F
|
617 |
+
type: chrf
|
618 |
+
value: 0.49524
|
619 |
+
- task:
|
620 |
+
name: Translation por-hne
|
621 |
+
type: translation
|
622 |
+
args: por-hne
|
623 |
+
dataset:
|
624 |
+
name: flores200-devtest
|
625 |
+
type: flores200-devtest
|
626 |
+
args: por-hne
|
627 |
+
metrics:
|
628 |
+
- name: BLEU
|
629 |
+
type: bleu
|
630 |
+
value: 13.9
|
631 |
+
- name: chr-F
|
632 |
+
type: chrf
|
633 |
+
value: 0.42269
|
634 |
+
- task:
|
635 |
+
name: Translation por-mag
|
636 |
+
type: translation
|
637 |
+
args: por-mag
|
638 |
+
dataset:
|
639 |
+
name: flores200-devtest
|
640 |
+
type: flores200-devtest
|
641 |
+
args: por-mag
|
642 |
+
metrics:
|
643 |
+
- name: BLEU
|
644 |
+
type: bleu
|
645 |
+
value: 15.0
|
646 |
+
- name: chr-F
|
647 |
+
type: chrf
|
648 |
+
value: 0.42753
|
649 |
+
- task:
|
650 |
+
name: Translation por-pan
|
651 |
+
type: translation
|
652 |
+
args: por-pan
|
653 |
+
dataset:
|
654 |
+
name: flores200-devtest
|
655 |
+
type: flores200-devtest
|
656 |
+
args: por-pan
|
657 |
+
metrics:
|
658 |
+
- name: BLEU
|
659 |
+
type: bleu
|
660 |
+
value: 10.7
|
661 |
+
- name: chr-F
|
662 |
+
type: chrf
|
663 |
+
value: 0.36653
|
664 |
+
- task:
|
665 |
+
name: Translation por-pes
|
666 |
+
type: translation
|
667 |
+
args: por-pes
|
668 |
+
dataset:
|
669 |
+
name: flores200-devtest
|
670 |
+
type: flores200-devtest
|
671 |
+
args: por-pes
|
672 |
+
metrics:
|
673 |
+
- name: BLEU
|
674 |
+
type: bleu
|
675 |
+
value: 15.4
|
676 |
+
- name: chr-F
|
677 |
+
type: chrf
|
678 |
+
value: 0.43194
|
679 |
+
- task:
|
680 |
+
name: Translation por-prs
|
681 |
+
type: translation
|
682 |
+
args: por-prs
|
683 |
+
dataset:
|
684 |
+
name: flores200-devtest
|
685 |
+
type: flores200-devtest
|
686 |
+
args: por-prs
|
687 |
+
metrics:
|
688 |
+
- name: BLEU
|
689 |
+
type: bleu
|
690 |
+
value: 12.1
|
691 |
+
- name: chr-F
|
692 |
+
type: chrf
|
693 |
+
value: 0.36411
|
694 |
+
- task:
|
695 |
+
name: Translation por-tgk
|
696 |
+
type: translation
|
697 |
+
args: por-tgk
|
698 |
+
dataset:
|
699 |
+
name: flores200-devtest
|
700 |
+
type: flores200-devtest
|
701 |
+
args: por-tgk
|
702 |
+
metrics:
|
703 |
+
- name: BLEU
|
704 |
+
type: bleu
|
705 |
+
value: 13.2
|
706 |
+
- name: chr-F
|
707 |
+
type: chrf
|
708 |
+
value: 0.41860
|
709 |
+
- task:
|
710 |
+
name: Translation por-urd
|
711 |
+
type: translation
|
712 |
+
args: por-urd
|
713 |
+
dataset:
|
714 |
+
name: flores200-devtest
|
715 |
+
type: flores200-devtest
|
716 |
+
args: por-urd
|
717 |
+
metrics:
|
718 |
+
- name: BLEU
|
719 |
+
type: bleu
|
720 |
+
value: 14.8
|
721 |
+
- name: chr-F
|
722 |
+
type: chrf
|
723 |
+
value: 0.41799
|
724 |
+
- task:
|
725 |
+
name: Translation spa-hin
|
726 |
+
type: translation
|
727 |
+
args: spa-hin
|
728 |
+
dataset:
|
729 |
+
name: flores200-devtest
|
730 |
+
type: flores200-devtest
|
731 |
+
args: spa-hin
|
732 |
+
metrics:
|
733 |
+
- name: BLEU
|
734 |
+
type: bleu
|
735 |
+
value: 16.4
|
736 |
+
- name: chr-F
|
737 |
+
type: chrf
|
738 |
+
value: 0.43777
|
739 |
+
- task:
|
740 |
+
name: Translation spa-hne
|
741 |
+
type: translation
|
742 |
+
args: spa-hne
|
743 |
+
dataset:
|
744 |
+
name: flores200-devtest
|
745 |
+
type: flores200-devtest
|
746 |
+
args: spa-hne
|
747 |
+
metrics:
|
748 |
+
- name: BLEU
|
749 |
+
type: bleu
|
750 |
+
value: 11.2
|
751 |
+
- name: chr-F
|
752 |
+
type: chrf
|
753 |
+
value: 0.39492
|
754 |
+
- task:
|
755 |
+
name: Translation spa-mag
|
756 |
+
type: translation
|
757 |
+
args: spa-mag
|
758 |
+
dataset:
|
759 |
+
name: flores200-devtest
|
760 |
+
type: flores200-devtest
|
761 |
+
args: spa-mag
|
762 |
+
metrics:
|
763 |
+
- name: BLEU
|
764 |
+
type: bleu
|
765 |
+
value: 11.4
|
766 |
+
- name: chr-F
|
767 |
+
type: chrf
|
768 |
+
value: 0.39882
|
769 |
+
- task:
|
770 |
+
name: Translation spa-pes
|
771 |
+
type: translation
|
772 |
+
args: spa-pes
|
773 |
+
dataset:
|
774 |
+
name: flores200-devtest
|
775 |
+
type: flores200-devtest
|
776 |
+
args: spa-pes
|
777 |
+
metrics:
|
778 |
+
- name: BLEU
|
779 |
+
type: bleu
|
780 |
+
value: 12.2
|
781 |
+
- name: chr-F
|
782 |
+
type: chrf
|
783 |
+
value: 0.40856
|
784 |
+
- task:
|
785 |
+
name: Translation spa-prs
|
786 |
+
type: translation
|
787 |
+
args: spa-prs
|
788 |
+
dataset:
|
789 |
+
name: flores200-devtest
|
790 |
+
type: flores200-devtest
|
791 |
+
args: spa-prs
|
792 |
+
metrics:
|
793 |
+
- name: BLEU
|
794 |
+
type: bleu
|
795 |
+
value: 12.8
|
796 |
+
- name: chr-F
|
797 |
+
type: chrf
|
798 |
+
value: 0.40361
|
799 |
+
- task:
|
800 |
+
name: Translation spa-tgk
|
801 |
+
type: translation
|
802 |
+
args: spa-tgk
|
803 |
+
dataset:
|
804 |
+
name: flores200-devtest
|
805 |
+
type: flores200-devtest
|
806 |
+
args: spa-tgk
|
807 |
+
metrics:
|
808 |
+
- name: BLEU
|
809 |
+
type: bleu
|
810 |
+
value: 10.8
|
811 |
+
- name: chr-F
|
812 |
+
type: chrf
|
813 |
+
value: 0.40100
|
814 |
+
- task:
|
815 |
+
name: Translation spa-urd
|
816 |
+
type: translation
|
817 |
+
args: spa-urd
|
818 |
+
dataset:
|
819 |
+
name: flores200-devtest
|
820 |
+
type: flores200-devtest
|
821 |
+
args: spa-urd
|
822 |
+
metrics:
|
823 |
+
- name: BLEU
|
824 |
+
type: bleu
|
825 |
+
value: 10.9
|
826 |
+
- name: chr-F
|
827 |
+
type: chrf
|
828 |
+
value: 0.38539
|
829 |
+
- task:
|
830 |
+
name: Translation deu-pan
|
831 |
+
type: translation
|
832 |
+
args: deu-pan
|
833 |
+
dataset:
|
834 |
+
name: flores101-devtest
|
835 |
+
type: flores_101
|
836 |
+
args: deu pan devtest
|
837 |
+
metrics:
|
838 |
+
- name: BLEU
|
839 |
+
type: bleu
|
840 |
+
value: 10.9
|
841 |
+
- name: chr-F
|
842 |
+
type: chrf
|
843 |
+
value: 0.36883
|
844 |
+
- task:
|
845 |
+
name: Translation eng-ben
|
846 |
+
type: translation
|
847 |
+
args: eng-ben
|
848 |
+
dataset:
|
849 |
+
name: flores101-devtest
|
850 |
+
type: flores_101
|
851 |
+
args: eng ben devtest
|
852 |
+
metrics:
|
853 |
+
- name: BLEU
|
854 |
+
type: bleu
|
855 |
+
value: 17.0
|
856 |
+
- name: chr-F
|
857 |
+
type: chrf
|
858 |
+
value: 0.51055
|
859 |
+
- task:
|
860 |
+
name: Translation eng-guj
|
861 |
+
type: translation
|
862 |
+
args: eng-guj
|
863 |
+
dataset:
|
864 |
+
name: flores101-devtest
|
865 |
+
type: flores_101
|
866 |
+
args: eng guj devtest
|
867 |
+
metrics:
|
868 |
+
- name: BLEU
|
869 |
+
type: bleu
|
870 |
+
value: 22.3
|
871 |
+
- name: chr-F
|
872 |
+
type: chrf
|
873 |
+
value: 0.53972
|
874 |
+
- task:
|
875 |
+
name: Translation eng-hin
|
876 |
+
type: translation
|
877 |
+
args: eng-hin
|
878 |
+
dataset:
|
879 |
+
name: flores101-devtest
|
880 |
+
type: flores_101
|
881 |
+
args: eng hin devtest
|
882 |
+
metrics:
|
883 |
+
- name: BLEU
|
884 |
+
type: bleu
|
885 |
+
value: 33.4
|
886 |
+
- name: chr-F
|
887 |
+
type: chrf
|
888 |
+
value: 0.57980
|
889 |
+
- task:
|
890 |
+
name: Translation eng-mar
|
891 |
+
type: translation
|
892 |
+
args: eng-mar
|
893 |
+
dataset:
|
894 |
+
name: flores101-devtest
|
895 |
+
type: flores_101
|
896 |
+
args: eng mar devtest
|
897 |
+
metrics:
|
898 |
+
- name: BLEU
|
899 |
+
type: bleu
|
900 |
+
value: 14.3
|
901 |
+
- name: chr-F
|
902 |
+
type: chrf
|
903 |
+
value: 0.48206
|
904 |
+
- task:
|
905 |
+
name: Translation eng-pus
|
906 |
+
type: translation
|
907 |
+
args: eng-pus
|
908 |
+
dataset:
|
909 |
+
name: flores101-devtest
|
910 |
+
type: flores_101
|
911 |
+
args: eng pus devtest
|
912 |
+
metrics:
|
913 |
+
- name: BLEU
|
914 |
+
type: bleu
|
915 |
+
value: 11.9
|
916 |
+
- name: chr-F
|
917 |
+
type: chrf
|
918 |
+
value: 0.37264
|
919 |
+
- task:
|
920 |
+
name: Translation eng-urd
|
921 |
+
type: translation
|
922 |
+
args: eng-urd
|
923 |
+
dataset:
|
924 |
+
name: flores101-devtest
|
925 |
+
type: flores_101
|
926 |
+
args: eng urd devtest
|
927 |
+
metrics:
|
928 |
+
- name: BLEU
|
929 |
+
type: bleu
|
930 |
+
value: 20.5
|
931 |
+
- name: chr-F
|
932 |
+
type: chrf
|
933 |
+
value: 0.48050
|
934 |
+
- task:
|
935 |
+
name: Translation fra-ben
|
936 |
+
type: translation
|
937 |
+
args: fra-ben
|
938 |
+
dataset:
|
939 |
+
name: flores101-devtest
|
940 |
+
type: flores_101
|
941 |
+
args: fra ben devtest
|
942 |
+
metrics:
|
943 |
+
- name: BLEU
|
944 |
+
type: bleu
|
945 |
+
value: 10.9
|
946 |
+
- name: chr-F
|
947 |
+
type: chrf
|
948 |
+
value: 0.43806
|
949 |
+
- task:
|
950 |
+
name: Translation fra-pan
|
951 |
+
type: translation
|
952 |
+
args: fra-pan
|
953 |
+
dataset:
|
954 |
+
name: flores101-devtest
|
955 |
+
type: flores_101
|
956 |
+
args: fra pan devtest
|
957 |
+
metrics:
|
958 |
+
- name: BLEU
|
959 |
+
type: bleu
|
960 |
+
value: 11.0
|
961 |
+
- name: chr-F
|
962 |
+
type: chrf
|
963 |
+
value: 0.37066
|
964 |
+
- task:
|
965 |
+
name: Translation por-ben
|
966 |
+
type: translation
|
967 |
+
args: por-ben
|
968 |
+
dataset:
|
969 |
+
name: flores101-devtest
|
970 |
+
type: flores_101
|
971 |
+
args: por ben devtest
|
972 |
+
metrics:
|
973 |
+
- name: BLEU
|
974 |
+
type: bleu
|
975 |
+
value: 10.0
|
976 |
+
- name: chr-F
|
977 |
+
type: chrf
|
978 |
+
value: 0.42730
|
979 |
+
- task:
|
980 |
+
name: Translation por-pan
|
981 |
+
type: translation
|
982 |
+
args: por-pan
|
983 |
+
dataset:
|
984 |
+
name: flores101-devtest
|
985 |
+
type: flores_101
|
986 |
+
args: por pan devtest
|
987 |
+
metrics:
|
988 |
+
- name: BLEU
|
989 |
+
type: bleu
|
990 |
+
value: 10.7
|
991 |
+
- name: chr-F
|
992 |
+
type: chrf
|
993 |
+
value: 0.36551
|
994 |
+
- task:
|
995 |
+
name: Translation spa-hin
|
996 |
+
type: translation
|
997 |
+
args: spa-hin
|
998 |
+
dataset:
|
999 |
+
name: flores101-devtest
|
1000 |
+
type: flores_101
|
1001 |
+
args: spa hin devtest
|
1002 |
+
metrics:
|
1003 |
+
- name: BLEU
|
1004 |
+
type: bleu
|
1005 |
+
value: 16.0
|
1006 |
+
- name: chr-F
|
1007 |
+
type: chrf
|
1008 |
+
value: 0.43371
|
1009 |
+
- task:
|
1010 |
+
name: Translation spa-tgk
|
1011 |
+
type: translation
|
1012 |
+
args: spa-tgk
|
1013 |
+
dataset:
|
1014 |
+
name: flores101-devtest
|
1015 |
+
type: flores_101
|
1016 |
+
args: spa tgk devtest
|
1017 |
+
metrics:
|
1018 |
+
- name: BLEU
|
1019 |
+
type: bleu
|
1020 |
+
value: 10.6
|
1021 |
+
- name: chr-F
|
1022 |
+
type: chrf
|
1023 |
+
value: 0.39762
|
1024 |
+
- task:
|
1025 |
+
name: Translation deu-fas
|
1026 |
+
type: translation
|
1027 |
+
args: deu-fas
|
1028 |
+
dataset:
|
1029 |
+
name: ntrex128
|
1030 |
+
type: ntrex128
|
1031 |
+
args: deu-fas
|
1032 |
+
metrics:
|
1033 |
+
- name: BLEU
|
1034 |
+
type: bleu
|
1035 |
+
value: 13.8
|
1036 |
+
- name: chr-F
|
1037 |
+
type: chrf
|
1038 |
+
value: 0.41469
|
1039 |
+
- task:
|
1040 |
+
name: Translation deu-hin
|
1041 |
+
type: translation
|
1042 |
+
args: deu-hin
|
1043 |
+
dataset:
|
1044 |
+
name: ntrex128
|
1045 |
+
type: ntrex128
|
1046 |
+
args: deu-hin
|
1047 |
+
metrics:
|
1048 |
+
- name: BLEU
|
1049 |
+
type: bleu
|
1050 |
+
value: 16.8
|
1051 |
+
- name: chr-F
|
1052 |
+
type: chrf
|
1053 |
+
value: 0.42940
|
1054 |
+
- task:
|
1055 |
+
name: Translation deu-pan
|
1056 |
+
type: translation
|
1057 |
+
args: deu-pan
|
1058 |
+
dataset:
|
1059 |
+
name: ntrex128
|
1060 |
+
type: ntrex128
|
1061 |
+
args: deu-pan
|
1062 |
+
metrics:
|
1063 |
+
- name: BLEU
|
1064 |
+
type: bleu
|
1065 |
+
value: 11.0
|
1066 |
+
- name: chr-F
|
1067 |
+
type: chrf
|
1068 |
+
value: 0.36776
|
1069 |
+
- task:
|
1070 |
+
name: Translation deu-urd
|
1071 |
+
type: translation
|
1072 |
+
args: deu-urd
|
1073 |
+
dataset:
|
1074 |
+
name: ntrex128
|
1075 |
+
type: ntrex128
|
1076 |
+
args: deu-urd
|
1077 |
+
metrics:
|
1078 |
+
- name: BLEU
|
1079 |
+
type: bleu
|
1080 |
+
value: 14.5
|
1081 |
+
- name: chr-F
|
1082 |
+
type: chrf
|
1083 |
+
value: 0.41881
|
1084 |
+
- task:
|
1085 |
+
name: Translation eng-ben
|
1086 |
+
type: translation
|
1087 |
+
args: eng-ben
|
1088 |
+
dataset:
|
1089 |
+
name: ntrex128
|
1090 |
+
type: ntrex128
|
1091 |
+
args: eng-ben
|
1092 |
+
metrics:
|
1093 |
+
- name: BLEU
|
1094 |
+
type: bleu
|
1095 |
+
value: 16.6
|
1096 |
+
- name: chr-F
|
1097 |
+
type: chrf
|
1098 |
+
value: 0.51555
|
1099 |
+
- task:
|
1100 |
+
name: Translation eng-fas
|
1101 |
+
type: translation
|
1102 |
+
args: eng-fas
|
1103 |
+
dataset:
|
1104 |
+
name: ntrex128
|
1105 |
+
type: ntrex128
|
1106 |
+
args: eng-fas
|
1107 |
+
metrics:
|
1108 |
+
- name: BLEU
|
1109 |
+
type: bleu
|
1110 |
+
value: 19.7
|
1111 |
+
- name: chr-F
|
1112 |
+
type: chrf
|
1113 |
+
value: 0.46895
|
1114 |
+
- task:
|
1115 |
+
name: Translation eng-guj
|
1116 |
+
type: translation
|
1117 |
+
args: eng-guj
|
1118 |
+
dataset:
|
1119 |
+
name: ntrex128
|
1120 |
+
type: ntrex128
|
1121 |
+
args: eng-guj
|
1122 |
+
metrics:
|
1123 |
+
- name: BLEU
|
1124 |
+
type: bleu
|
1125 |
+
value: 17.1
|
1126 |
+
- name: chr-F
|
1127 |
+
type: chrf
|
1128 |
+
value: 0.48990
|
1129 |
+
- task:
|
1130 |
+
name: Translation eng-hin
|
1131 |
+
type: translation
|
1132 |
+
args: eng-hin
|
1133 |
+
dataset:
|
1134 |
+
name: ntrex128
|
1135 |
+
type: ntrex128
|
1136 |
+
args: eng-hin
|
1137 |
+
metrics:
|
1138 |
+
- name: BLEU
|
1139 |
+
type: bleu
|
1140 |
+
value: 26.9
|
1141 |
+
- name: chr-F
|
1142 |
+
type: chrf
|
1143 |
+
value: 0.52307
|
1144 |
+
- task:
|
1145 |
+
name: Translation eng-mar
|
1146 |
+
type: translation
|
1147 |
+
args: eng-mar
|
1148 |
+
dataset:
|
1149 |
+
name: ntrex128
|
1150 |
+
type: ntrex128
|
1151 |
+
args: eng-mar
|
1152 |
+
metrics:
|
1153 |
+
- name: BLEU
|
1154 |
+
type: bleu
|
1155 |
+
value: 10.4
|
1156 |
+
- name: chr-F
|
1157 |
+
type: chrf
|
1158 |
+
value: 0.44580
|
1159 |
+
- task:
|
1160 |
+
name: Translation eng-pan
|
1161 |
+
type: translation
|
1162 |
+
args: eng-pan
|
1163 |
+
dataset:
|
1164 |
+
name: ntrex128
|
1165 |
+
type: ntrex128
|
1166 |
+
args: eng-pan
|
1167 |
+
metrics:
|
1168 |
+
- name: BLEU
|
1169 |
+
type: bleu
|
1170 |
+
value: 19.6
|
1171 |
+
- name: chr-F
|
1172 |
+
type: chrf
|
1173 |
+
value: 0.46141
|
1174 |
+
- task:
|
1175 |
+
name: Translation eng-prs
|
1176 |
+
type: translation
|
1177 |
+
args: eng-prs
|
1178 |
+
dataset:
|
1179 |
+
name: ntrex128
|
1180 |
+
type: ntrex128
|
1181 |
+
args: eng-prs
|
1182 |
+
metrics:
|
1183 |
+
- name: BLEU
|
1184 |
+
type: bleu
|
1185 |
+
value: 12.9
|
1186 |
+
- name: chr-F
|
1187 |
+
type: chrf
|
1188 |
+
value: 0.39651
|
1189 |
+
- task:
|
1190 |
+
name: Translation eng-tgk_Cyrl
|
1191 |
+
type: translation
|
1192 |
+
args: eng-tgk_Cyrl
|
1193 |
+
dataset:
|
1194 |
+
name: ntrex128
|
1195 |
+
type: ntrex128
|
1196 |
+
args: eng-tgk_Cyrl
|
1197 |
+
metrics:
|
1198 |
+
- name: BLEU
|
1199 |
+
type: bleu
|
1200 |
+
value: 11.3
|
1201 |
+
- name: chr-F
|
1202 |
+
type: chrf
|
1203 |
+
value: 0.38524
|
1204 |
+
- task:
|
1205 |
+
name: Translation eng-urd
|
1206 |
+
type: translation
|
1207 |
+
args: eng-urd
|
1208 |
+
dataset:
|
1209 |
+
name: ntrex128
|
1210 |
+
type: ntrex128
|
1211 |
+
args: eng-urd
|
1212 |
+
metrics:
|
1213 |
+
- name: BLEU
|
1214 |
+
type: bleu
|
1215 |
+
value: 22.1
|
1216 |
+
- name: chr-F
|
1217 |
+
type: chrf
|
1218 |
+
value: 0.49646
|
1219 |
+
- task:
|
1220 |
+
name: Translation fra-fas
|
1221 |
+
type: translation
|
1222 |
+
args: fra-fas
|
1223 |
+
dataset:
|
1224 |
+
name: ntrex128
|
1225 |
+
type: ntrex128
|
1226 |
+
args: fra-fas
|
1227 |
+
metrics:
|
1228 |
+
- name: BLEU
|
1229 |
+
type: bleu
|
1230 |
+
value: 13.8
|
1231 |
+
- name: chr-F
|
1232 |
+
type: chrf
|
1233 |
+
value: 0.41282
|
1234 |
+
- task:
|
1235 |
+
name: Translation fra-hin
|
1236 |
+
type: translation
|
1237 |
+
args: fra-hin
|
1238 |
+
dataset:
|
1239 |
+
name: ntrex128
|
1240 |
+
type: ntrex128
|
1241 |
+
args: fra-hin
|
1242 |
+
metrics:
|
1243 |
+
- name: BLEU
|
1244 |
+
type: bleu
|
1245 |
+
value: 17.1
|
1246 |
+
- name: chr-F
|
1247 |
+
type: chrf
|
1248 |
+
value: 0.42475
|
1249 |
+
- task:
|
1250 |
+
name: Translation fra-pan
|
1251 |
+
type: translation
|
1252 |
+
args: fra-pan
|
1253 |
+
dataset:
|
1254 |
+
name: ntrex128
|
1255 |
+
type: ntrex128
|
1256 |
+
args: fra-pan
|
1257 |
+
metrics:
|
1258 |
+
- name: BLEU
|
1259 |
+
type: bleu
|
1260 |
+
value: 10.2
|
1261 |
+
- name: chr-F
|
1262 |
+
type: chrf
|
1263 |
+
value: 0.36120
|
1264 |
+
- task:
|
1265 |
+
name: Translation fra-urd
|
1266 |
+
type: translation
|
1267 |
+
args: fra-urd
|
1268 |
+
dataset:
|
1269 |
+
name: ntrex128
|
1270 |
+
type: ntrex128
|
1271 |
+
args: fra-urd
|
1272 |
+
metrics:
|
1273 |
+
- name: BLEU
|
1274 |
+
type: bleu
|
1275 |
+
value: 14.8
|
1276 |
+
- name: chr-F
|
1277 |
+
type: chrf
|
1278 |
+
value: 0.41536
|
1279 |
+
- task:
|
1280 |
+
name: Translation por-fas
|
1281 |
+
type: translation
|
1282 |
+
args: por-fas
|
1283 |
+
dataset:
|
1284 |
+
name: ntrex128
|
1285 |
+
type: ntrex128
|
1286 |
+
args: por-fas
|
1287 |
+
metrics:
|
1288 |
+
- name: BLEU
|
1289 |
+
type: bleu
|
1290 |
+
value: 14.4
|
1291 |
+
- name: chr-F
|
1292 |
+
type: chrf
|
1293 |
+
value: 0.42010
|
1294 |
+
- task:
|
1295 |
+
name: Translation por-hin
|
1296 |
+
type: translation
|
1297 |
+
args: por-hin
|
1298 |
+
dataset:
|
1299 |
+
name: ntrex128
|
1300 |
+
type: ntrex128
|
1301 |
+
args: por-hin
|
1302 |
+
metrics:
|
1303 |
+
- name: BLEU
|
1304 |
+
type: bleu
|
1305 |
+
value: 17.6
|
1306 |
+
- name: chr-F
|
1307 |
+
type: chrf
|
1308 |
+
value: 0.43275
|
1309 |
+
- task:
|
1310 |
+
name: Translation por-pan
|
1311 |
+
type: translation
|
1312 |
+
args: por-pan
|
1313 |
+
dataset:
|
1314 |
+
name: ntrex128
|
1315 |
+
type: ntrex128
|
1316 |
+
args: por-pan
|
1317 |
+
metrics:
|
1318 |
+
- name: BLEU
|
1319 |
+
type: bleu
|
1320 |
+
value: 10.6
|
1321 |
+
- name: chr-F
|
1322 |
+
type: chrf
|
1323 |
+
value: 0.36360
|
1324 |
+
- task:
|
1325 |
+
name: Translation por-urd
|
1326 |
+
type: translation
|
1327 |
+
args: por-urd
|
1328 |
+
dataset:
|
1329 |
+
name: ntrex128
|
1330 |
+
type: ntrex128
|
1331 |
+
args: por-urd
|
1332 |
+
metrics:
|
1333 |
+
- name: BLEU
|
1334 |
+
type: bleu
|
1335 |
+
value: 15.2
|
1336 |
+
- name: chr-F
|
1337 |
+
type: chrf
|
1338 |
+
value: 0.42484
|
1339 |
+
- task:
|
1340 |
+
name: Translation spa-ben
|
1341 |
+
type: translation
|
1342 |
+
args: spa-ben
|
1343 |
+
dataset:
|
1344 |
+
name: ntrex128
|
1345 |
+
type: ntrex128
|
1346 |
+
args: spa-ben
|
1347 |
+
metrics:
|
1348 |
+
- name: BLEU
|
1349 |
+
type: bleu
|
1350 |
+
value: 10.3
|
1351 |
+
- name: chr-F
|
1352 |
+
type: chrf
|
1353 |
+
value: 0.44905
|
1354 |
+
- task:
|
1355 |
+
name: Translation spa-fas
|
1356 |
+
type: translation
|
1357 |
+
args: spa-fas
|
1358 |
+
dataset:
|
1359 |
+
name: ntrex128
|
1360 |
+
type: ntrex128
|
1361 |
+
args: spa-fas
|
1362 |
+
metrics:
|
1363 |
+
- name: BLEU
|
1364 |
+
type: bleu
|
1365 |
+
value: 14.1
|
1366 |
+
- name: chr-F
|
1367 |
+
type: chrf
|
1368 |
+
value: 0.42207
|
1369 |
+
- task:
|
1370 |
+
name: Translation spa-hin
|
1371 |
+
type: translation
|
1372 |
+
args: spa-hin
|
1373 |
+
dataset:
|
1374 |
+
name: ntrex128
|
1375 |
+
type: ntrex128
|
1376 |
+
args: spa-hin
|
1377 |
+
metrics:
|
1378 |
+
- name: BLEU
|
1379 |
+
type: bleu
|
1380 |
+
value: 17.6
|
1381 |
+
- name: chr-F
|
1382 |
+
type: chrf
|
1383 |
+
value: 0.43380
|
1384 |
+
- task:
|
1385 |
+
name: Translation spa-pan
|
1386 |
+
type: translation
|
1387 |
+
args: spa-pan
|
1388 |
+
dataset:
|
1389 |
+
name: ntrex128
|
1390 |
+
type: ntrex128
|
1391 |
+
args: spa-pan
|
1392 |
+
metrics:
|
1393 |
+
- name: BLEU
|
1394 |
+
type: bleu
|
1395 |
+
value: 11.1
|
1396 |
+
- name: chr-F
|
1397 |
+
type: chrf
|
1398 |
+
value: 0.37361
|
1399 |
+
- task:
|
1400 |
+
name: Translation spa-prs
|
1401 |
+
type: translation
|
1402 |
+
args: spa-prs
|
1403 |
+
dataset:
|
1404 |
+
name: ntrex128
|
1405 |
+
type: ntrex128
|
1406 |
+
args: spa-prs
|
1407 |
+
metrics:
|
1408 |
+
- name: BLEU
|
1409 |
+
type: bleu
|
1410 |
+
value: 10.3
|
1411 |
+
- name: chr-F
|
1412 |
+
type: chrf
|
1413 |
+
value: 0.37448
|
1414 |
+
- task:
|
1415 |
+
name: Translation spa-urd
|
1416 |
+
type: translation
|
1417 |
+
args: spa-urd
|
1418 |
+
dataset:
|
1419 |
+
name: ntrex128
|
1420 |
+
type: ntrex128
|
1421 |
+
args: spa-urd
|
1422 |
+
metrics:
|
1423 |
+
- name: BLEU
|
1424 |
+
type: bleu
|
1425 |
+
value: 15.0
|
1426 |
+
- name: chr-F
|
1427 |
+
type: chrf
|
1428 |
+
value: 0.42434
|
1429 |
+
- task:
|
1430 |
+
name: Translation deu-fas
|
1431 |
+
type: translation
|
1432 |
+
args: deu-fas
|
1433 |
+
dataset:
|
1434 |
+
name: tatoeba-test-v2021-08-07
|
1435 |
+
type: tatoeba_mt
|
1436 |
+
args: deu-fas
|
1437 |
+
metrics:
|
1438 |
+
- name: BLEU
|
1439 |
+
type: bleu
|
1440 |
+
value: 20.3
|
1441 |
+
- name: chr-F
|
1442 |
+
type: chrf
|
1443 |
+
value: 0.45763
|
1444 |
+
- task:
|
1445 |
+
name: Translation eng-hin
|
1446 |
+
type: translation
|
1447 |
+
args: eng-hin
|
1448 |
+
dataset:
|
1449 |
+
name: tatoeba-test-v2021-08-07
|
1450 |
+
type: tatoeba_mt
|
1451 |
+
args: eng-hin
|
1452 |
+
metrics:
|
1453 |
+
- name: BLEU
|
1454 |
+
type: bleu
|
1455 |
+
value: 28.4
|
1456 |
+
- name: chr-F
|
1457 |
+
type: chrf
|
1458 |
+
value: 0.52525
|
1459 |
+
- task:
|
1460 |
+
name: Translation eng-mar
|
1461 |
+
type: translation
|
1462 |
+
args: eng-mar
|
1463 |
+
dataset:
|
1464 |
+
name: tatoeba-test-v2021-08-07
|
1465 |
+
type: tatoeba_mt
|
1466 |
+
args: eng-mar
|
1467 |
+
metrics:
|
1468 |
+
- name: BLEU
|
1469 |
+
type: bleu
|
1470 |
+
value: 24.4
|
1471 |
+
- name: chr-F
|
1472 |
+
type: chrf
|
1473 |
+
value: 0.52549
|
1474 |
+
- task:
|
1475 |
+
name: Translation eng-ben
|
1476 |
+
type: translation
|
1477 |
+
args: eng-ben
|
1478 |
+
dataset:
|
1479 |
+
name: tico19-test
|
1480 |
+
type: tico19-test
|
1481 |
+
args: eng-ben
|
1482 |
+
metrics:
|
1483 |
+
- name: BLEU
|
1484 |
+
type: bleu
|
1485 |
+
value: 17.9
|
1486 |
+
- name: chr-F
|
1487 |
+
type: chrf
|
1488 |
+
value: 0.51563
|
1489 |
+
- task:
|
1490 |
+
name: Translation eng-fas
|
1491 |
+
type: translation
|
1492 |
+
args: eng-fas
|
1493 |
+
dataset:
|
1494 |
+
name: tico19-test
|
1495 |
+
type: tico19-test
|
1496 |
+
args: eng-fas
|
1497 |
+
metrics:
|
1498 |
+
- name: BLEU
|
1499 |
+
type: bleu
|
1500 |
+
value: 25.8
|
1501 |
+
- name: chr-F
|
1502 |
+
type: chrf
|
1503 |
+
value: 0.53182
|
1504 |
+
- task:
|
1505 |
+
name: Translation eng-hin
|
1506 |
+
type: translation
|
1507 |
+
args: eng-hin
|
1508 |
+
dataset:
|
1509 |
+
name: tico19-test
|
1510 |
+
type: tico19-test
|
1511 |
+
args: eng-hin
|
1512 |
+
metrics:
|
1513 |
+
- name: BLEU
|
1514 |
+
type: bleu
|
1515 |
+
value: 41.6
|
1516 |
+
- name: chr-F
|
1517 |
+
type: chrf
|
1518 |
+
value: 0.63128
|
1519 |
+
- task:
|
1520 |
+
name: Translation eng-mar
|
1521 |
+
type: translation
|
1522 |
+
args: eng-mar
|
1523 |
+
dataset:
|
1524 |
+
name: tico19-test
|
1525 |
+
type: tico19-test
|
1526 |
+
args: eng-mar
|
1527 |
+
metrics:
|
1528 |
+
- name: BLEU
|
1529 |
+
type: bleu
|
1530 |
+
value: 12.9
|
1531 |
+
- name: chr-F
|
1532 |
+
type: chrf
|
1533 |
+
value: 0.45619
|
1534 |
+
- task:
|
1535 |
+
name: Translation eng-nep
|
1536 |
+
type: translation
|
1537 |
+
args: eng-nep
|
1538 |
+
dataset:
|
1539 |
+
name: tico19-test
|
1540 |
+
type: tico19-test
|
1541 |
+
args: eng-nep
|
1542 |
+
metrics:
|
1543 |
+
- name: BLEU
|
1544 |
+
type: bleu
|
1545 |
+
value: 17.6
|
1546 |
+
- name: chr-F
|
1547 |
+
type: chrf
|
1548 |
+
value: 0.53413
|
1549 |
+
- task:
|
1550 |
+
name: Translation eng-prs
|
1551 |
+
type: translation
|
1552 |
+
args: eng-prs
|
1553 |
+
dataset:
|
1554 |
+
name: tico19-test
|
1555 |
+
type: tico19-test
|
1556 |
+
args: eng-prs
|
1557 |
+
metrics:
|
1558 |
+
- name: BLEU
|
1559 |
+
type: bleu
|
1560 |
+
value: 17.3
|
1561 |
+
- name: chr-F
|
1562 |
+
type: chrf
|
1563 |
+
value: 0.44101
|
1564 |
+
- task:
|
1565 |
+
name: Translation eng-pus
|
1566 |
+
type: translation
|
1567 |
+
args: eng-pus
|
1568 |
+
dataset:
|
1569 |
+
name: tico19-test
|
1570 |
+
type: tico19-test
|
1571 |
+
args: eng-pus
|
1572 |
+
metrics:
|
1573 |
+
- name: BLEU
|
1574 |
+
type: bleu
|
1575 |
+
value: 20.5
|
1576 |
+
- name: chr-F
|
1577 |
+
type: chrf
|
1578 |
+
value: 0.47063
|
1579 |
+
- task:
|
1580 |
+
name: Translation eng-urd
|
1581 |
+
type: translation
|
1582 |
+
args: eng-urd
|
1583 |
+
dataset:
|
1584 |
+
name: tico19-test
|
1585 |
+
type: tico19-test
|
1586 |
+
args: eng-urd
|
1587 |
+
metrics:
|
1588 |
+
- name: BLEU
|
1589 |
+
type: bleu
|
1590 |
+
value: 22.0
|
1591 |
+
- name: chr-F
|
1592 |
+
type: chrf
|
1593 |
+
value: 0.51054
|
1594 |
+
- task:
|
1595 |
+
name: Translation fra-fas
|
1596 |
+
type: translation
|
1597 |
+
args: fra-fas
|
1598 |
+
dataset:
|
1599 |
+
name: tico19-test
|
1600 |
+
type: tico19-test
|
1601 |
+
args: fra-fas
|
1602 |
+
metrics:
|
1603 |
+
- name: BLEU
|
1604 |
+
type: bleu
|
1605 |
+
value: 17.9
|
1606 |
+
- name: chr-F
|
1607 |
+
type: chrf
|
1608 |
+
value: 0.43476
|
1609 |
+
- task:
|
1610 |
+
name: Translation fra-hin
|
1611 |
+
type: translation
|
1612 |
+
args: fra-hin
|
1613 |
+
dataset:
|
1614 |
+
name: tico19-test
|
1615 |
+
type: tico19-test
|
1616 |
+
args: fra-hin
|
1617 |
+
metrics:
|
1618 |
+
- name: BLEU
|
1619 |
+
type: bleu
|
1620 |
+
value: 25.6
|
1621 |
+
- name: chr-F
|
1622 |
+
type: chrf
|
1623 |
+
value: 0.48625
|
1624 |
+
- task:
|
1625 |
+
name: Translation fra-prs
|
1626 |
+
type: translation
|
1627 |
+
args: fra-prs
|
1628 |
+
dataset:
|
1629 |
+
name: tico19-test
|
1630 |
+
type: tico19-test
|
1631 |
+
args: fra-prs
|
1632 |
+
metrics:
|
1633 |
+
- name: BLEU
|
1634 |
+
type: bleu
|
1635 |
+
value: 11.6
|
1636 |
+
- name: chr-F
|
1637 |
+
type: chrf
|
1638 |
+
value: 0.36130
|
1639 |
+
- task:
|
1640 |
+
name: Translation fra-pus
|
1641 |
+
type: translation
|
1642 |
+
args: fra-pus
|
1643 |
+
dataset:
|
1644 |
+
name: tico19-test
|
1645 |
+
type: tico19-test
|
1646 |
+
args: fra-pus
|
1647 |
+
metrics:
|
1648 |
+
- name: BLEU
|
1649 |
+
type: bleu
|
1650 |
+
value: 12.7
|
1651 |
+
- name: chr-F
|
1652 |
+
type: chrf
|
1653 |
+
value: 0.37217
|
1654 |
+
- task:
|
1655 |
+
name: Translation fra-urd
|
1656 |
+
type: translation
|
1657 |
+
args: fra-urd
|
1658 |
+
dataset:
|
1659 |
+
name: tico19-test
|
1660 |
+
type: tico19-test
|
1661 |
+
args: fra-urd
|
1662 |
+
metrics:
|
1663 |
+
- name: BLEU
|
1664 |
+
type: bleu
|
1665 |
+
value: 14.4
|
1666 |
+
- name: chr-F
|
1667 |
+
type: chrf
|
1668 |
+
value: 0.40482
|
1669 |
+
- task:
|
1670 |
+
name: Translation por-ben
|
1671 |
+
type: translation
|
1672 |
+
args: por-ben
|
1673 |
+
dataset:
|
1674 |
+
name: tico19-test
|
1675 |
+
type: tico19-test
|
1676 |
+
args: por-ben
|
1677 |
+
metrics:
|
1678 |
+
- name: BLEU
|
1679 |
+
type: bleu
|
1680 |
+
value: 12.5
|
1681 |
+
- name: chr-F
|
1682 |
+
type: chrf
|
1683 |
+
value: 0.45814
|
1684 |
+
- task:
|
1685 |
+
name: Translation por-fas
|
1686 |
+
type: translation
|
1687 |
+
args: por-fas
|
1688 |
+
dataset:
|
1689 |
+
name: tico19-test
|
1690 |
+
type: tico19-test
|
1691 |
+
args: por-fas
|
1692 |
+
metrics:
|
1693 |
+
- name: BLEU
|
1694 |
+
type: bleu
|
1695 |
+
value: 21.3
|
1696 |
+
- name: chr-F
|
1697 |
+
type: chrf
|
1698 |
+
value: 0.49181
|
1699 |
+
- task:
|
1700 |
+
name: Translation por-hin
|
1701 |
+
type: translation
|
1702 |
+
args: por-hin
|
1703 |
+
dataset:
|
1704 |
+
name: tico19-test
|
1705 |
+
type: tico19-test
|
1706 |
+
args: por-hin
|
1707 |
+
metrics:
|
1708 |
+
- name: BLEU
|
1709 |
+
type: bleu
|
1710 |
+
value: 31.1
|
1711 |
+
- name: chr-F
|
1712 |
+
type: chrf
|
1713 |
+
value: 0.55759
|
1714 |
+
- task:
|
1715 |
+
name: Translation por-nep
|
1716 |
+
type: translation
|
1717 |
+
args: por-nep
|
1718 |
+
dataset:
|
1719 |
+
name: tico19-test
|
1720 |
+
type: tico19-test
|
1721 |
+
args: por-nep
|
1722 |
+
metrics:
|
1723 |
+
- name: BLEU
|
1724 |
+
type: bleu
|
1725 |
+
value: 12.1
|
1726 |
+
- name: chr-F
|
1727 |
+
type: chrf
|
1728 |
+
value: 0.47378
|
1729 |
+
- task:
|
1730 |
+
name: Translation por-prs
|
1731 |
+
type: translation
|
1732 |
+
args: por-prs
|
1733 |
+
dataset:
|
1734 |
+
name: tico19-test
|
1735 |
+
type: tico19-test
|
1736 |
+
args: por-prs
|
1737 |
+
metrics:
|
1738 |
+
- name: BLEU
|
1739 |
+
type: bleu
|
1740 |
+
value: 12.1
|
1741 |
+
- name: chr-F
|
1742 |
+
type: chrf
|
1743 |
+
value: 0.38725
|
1744 |
+
- task:
|
1745 |
+
name: Translation por-pus
|
1746 |
+
type: translation
|
1747 |
+
args: por-pus
|
1748 |
+
dataset:
|
1749 |
+
name: tico19-test
|
1750 |
+
type: tico19-test
|
1751 |
+
args: por-pus
|
1752 |
+
metrics:
|
1753 |
+
- name: BLEU
|
1754 |
+
type: bleu
|
1755 |
+
value: 15.9
|
1756 |
+
- name: chr-F
|
1757 |
+
type: chrf
|
1758 |
+
value: 0.42496
|
1759 |
+
- task:
|
1760 |
+
name: Translation por-urd
|
1761 |
+
type: translation
|
1762 |
+
args: por-urd
|
1763 |
+
dataset:
|
1764 |
+
name: tico19-test
|
1765 |
+
type: tico19-test
|
1766 |
+
args: por-urd
|
1767 |
+
metrics:
|
1768 |
+
- name: BLEU
|
1769 |
+
type: bleu
|
1770 |
+
value: 16.6
|
1771 |
+
- name: chr-F
|
1772 |
+
type: chrf
|
1773 |
+
value: 0.45560
|
1774 |
+
- task:
|
1775 |
+
name: Translation spa-ben
|
1776 |
+
type: translation
|
1777 |
+
args: spa-ben
|
1778 |
+
dataset:
|
1779 |
+
name: tico19-test
|
1780 |
+
type: tico19-test
|
1781 |
+
args: spa-ben
|
1782 |
+
metrics:
|
1783 |
+
- name: BLEU
|
1784 |
+
type: bleu
|
1785 |
+
value: 12.7
|
1786 |
+
- name: chr-F
|
1787 |
+
type: chrf
|
1788 |
+
value: 0.45751
|
1789 |
+
- task:
|
1790 |
+
name: Translation spa-fas
|
1791 |
+
type: translation
|
1792 |
+
args: spa-fas
|
1793 |
+
dataset:
|
1794 |
+
name: tico19-test
|
1795 |
+
type: tico19-test
|
1796 |
+
args: spa-fas
|
1797 |
+
metrics:
|
1798 |
+
- name: BLEU
|
1799 |
+
type: bleu
|
1800 |
+
value: 21.0
|
1801 |
+
- name: chr-F
|
1802 |
+
type: chrf
|
1803 |
+
value: 0.48974
|
1804 |
+
- task:
|
1805 |
+
name: Translation spa-hin
|
1806 |
+
type: translation
|
1807 |
+
args: spa-hin
|
1808 |
+
dataset:
|
1809 |
+
name: tico19-test
|
1810 |
+
type: tico19-test
|
1811 |
+
args: spa-hin
|
1812 |
+
metrics:
|
1813 |
+
- name: BLEU
|
1814 |
+
type: bleu
|
1815 |
+
value: 30.9
|
1816 |
+
- name: chr-F
|
1817 |
+
type: chrf
|
1818 |
+
value: 0.55641
|
1819 |
+
- task:
|
1820 |
+
name: Translation spa-nep
|
1821 |
+
type: translation
|
1822 |
+
args: spa-nep
|
1823 |
+
dataset:
|
1824 |
+
name: tico19-test
|
1825 |
+
type: tico19-test
|
1826 |
+
args: spa-nep
|
1827 |
+
metrics:
|
1828 |
+
- name: BLEU
|
1829 |
+
type: bleu
|
1830 |
+
value: 12.1
|
1831 |
+
- name: chr-F
|
1832 |
+
type: chrf
|
1833 |
+
value: 0.47164
|
1834 |
+
- task:
|
1835 |
+
name: Translation spa-prs
|
1836 |
+
type: translation
|
1837 |
+
args: spa-prs
|
1838 |
+
dataset:
|
1839 |
+
name: tico19-test
|
1840 |
+
type: tico19-test
|
1841 |
+
args: spa-prs
|
1842 |
+
metrics:
|
1843 |
+
- name: BLEU
|
1844 |
+
type: bleu
|
1845 |
+
value: 14.3
|
1846 |
+
- name: chr-F
|
1847 |
+
type: chrf
|
1848 |
+
value: 0.41879
|
1849 |
+
- task:
|
1850 |
+
name: Translation spa-pus
|
1851 |
+
type: translation
|
1852 |
+
args: spa-pus
|
1853 |
+
dataset:
|
1854 |
+
name: tico19-test
|
1855 |
+
type: tico19-test
|
1856 |
+
args: spa-pus
|
1857 |
+
metrics:
|
1858 |
+
- name: BLEU
|
1859 |
+
type: bleu
|
1860 |
+
value: 15.1
|
1861 |
+
- name: chr-F
|
1862 |
+
type: chrf
|
1863 |
+
value: 0.41714
|
1864 |
+
- task:
|
1865 |
+
name: Translation spa-urd
|
1866 |
+
type: translation
|
1867 |
+
args: spa-urd
|
1868 |
+
dataset:
|
1869 |
+
name: tico19-test
|
1870 |
+
type: tico19-test
|
1871 |
+
args: spa-urd
|
1872 |
+
metrics:
|
1873 |
+
- name: BLEU
|
1874 |
+
type: bleu
|
1875 |
+
value: 15.3
|
1876 |
+
- name: chr-F
|
1877 |
+
type: chrf
|
1878 |
+
value: 0.44931
|
1879 |
+
- task:
|
1880 |
+
name: Translation eng-hin
|
1881 |
+
type: translation
|
1882 |
+
args: eng-hin
|
1883 |
+
dataset:
|
1884 |
+
name: newstest2014
|
1885 |
+
type: wmt-2014-news
|
1886 |
+
args: eng-hin
|
1887 |
+
metrics:
|
1888 |
+
- name: BLEU
|
1889 |
+
type: bleu
|
1890 |
+
value: 23.6
|
1891 |
+
- name: chr-F
|
1892 |
+
type: chrf
|
1893 |
+
value: 0.51249
|
1894 |
+
- task:
|
1895 |
+
name: Translation eng-guj
|
1896 |
+
type: translation
|
1897 |
+
args: eng-guj
|
1898 |
+
dataset:
|
1899 |
+
name: newstest2019
|
1900 |
+
type: wmt-2019-news
|
1901 |
+
args: eng-guj
|
1902 |
+
metrics:
|
1903 |
+
- name: BLEU
|
1904 |
+
type: bleu
|
1905 |
+
value: 25.5
|
1906 |
+
- name: chr-F
|
1907 |
+
type: chrf
|
1908 |
+
value: 0.57282
|
1909 |
+
---
|
1910 |
+
# opus-mt-tc-bible-big-deu_eng_fra_por_spa-iir
|
1911 |
+
|
1912 |
+
## Table of Contents
|
1913 |
+
- [Model Details](#model-details)
|
1914 |
+
- [Uses](#uses)
|
1915 |
+
- [Risks, Limitations and Biases](#risks-limitations-and-biases)
|
1916 |
+
- [How to Get Started With the Model](#how-to-get-started-with-the-model)
|
1917 |
+
- [Training](#training)
|
1918 |
+
- [Evaluation](#evaluation)
|
1919 |
+
- [Citation Information](#citation-information)
|
1920 |
+
- [Acknowledgements](#acknowledgements)
|
1921 |
+
|
1922 |
+
## Model Details
|
1923 |
+
|
1924 |
+
Neural machine translation model for translating from unknown (deu+eng+fra+por+spa) to Indo-Iranian languages (iir).
|
1925 |
+
|
1926 |
+
This model is part of the [OPUS-MT project](https://github.com/Helsinki-NLP/Opus-MT), an effort to make neural machine translation models widely available and accessible for many languages in the world. All models are originally trained using the amazing framework of [Marian NMT](https://marian-nmt.github.io/), an efficient NMT implementation written in pure C++. The models have been converted to pyTorch using the transformers library by huggingface. Training data is taken from [OPUS](https://opus.nlpl.eu/) and training pipelines use the procedures of [OPUS-MT-train](https://github.com/Helsinki-NLP/Opus-MT-train).
|
1927 |
+
**Model Description:**
|
1928 |
+
- **Developed by:** Language Technology Research Group at the University of Helsinki
|
1929 |
+
- **Model Type:** Translation (transformer-big)
|
1930 |
+
- **Release**: 2024-05-30
|
1931 |
+
- **License:** Apache-2.0
|
1932 |
+
- **Language(s):**
|
1933 |
+
- Source Language(s): deu eng fra por spa
|
1934 |
+
- Target Language(s): anp asm awa bal ben bho bpy ckb diq div dty fas gbm glk guj hif hin hne hns jdt kas kmr kok kur lah lrc mag mai mar mzn nep npi ori oss pal pan pes pli prs pus rhg rmy rom san sdh sin skr snd syl tgk tly urd zza
|
1935 |
+
- Valid Target Language Labels: >>aee<< >>aeq<< >>aiq<< >>anp<< >>anr<< >>ask<< >>asm<< >>atn<< >>avd<< >>ave<< >>awa<< >>bal<< >>bal_Latn<< >>bdv<< >>ben<< >>bfb<< >>bfy<< >>bfz<< >>bgc<< >>bgd<< >>bge<< >>bgw<< >>bha<< >>bhb<< >>bhd<< >>bhe<< >>bhh<< >>bhi<< >>bho<< >>bht<< >>bhu<< >>bjj<< >>bjm<< >>bkk<< >>bmj<< >>bns<< >>bpx<< >>bpy<< >>bqi<< >>bra<< >>bsg<< >>bsh<< >>btv<< >>ccp<< >>cdh<< >>cdi<< >>cdj<< >>cih<< >>ckb<< >>clh<< >>ctg<< >>dcc<< >>def<< >>deh<< >>dhn<< >>dho<< >>diq<< >>div<< >>dmk<< >>dml<< >>doi<< >>dry<< >>dty<< >>dub<< >>duh<< >>dwz<< >>emx<< >>esh<< >>fas<< >>fay<< >>gas<< >>gbk<< >>gbl<< >>gbm<< >>gbz<< >>gdx<< >>ggg<< >>ghr<< >>gig<< >>gjk<< >>glh<< >>glk<< >>goz<< >>gra<< >>guj<< >>gwc<< >>gwf<< >>gwt<< >>gzi<< >>hac<< >>haj<< >>haz<< >>hca<< >>hif<< >>hif_Latn<< >>hii<< >>hin<< >>hin_Latn<< >>hlb<< >>hne<< >>hns<< >>hrz<< >>isk<< >>jdg<< >>jdt<< >>jdt_Cyrl<< >>jml<< >>jnd<< >>jns<< >>jpr<< >>kas<< >>kas_Arab<< >>kas_Deva<< >>kbu<< >>keq<< >>key<< >>kfm<< >>kfr<< >>kfs<< >>kft<< >>kfu<< >>kfv<< >>kfx<< >>kfy<< >>kgn<< >>khn<< >>kho<< >>khw<< >>kjo<< >>kls<< >>kmr<< >>kok<< >>kra<< >>ksy<< >>ktl<< >>kur<< >>kur_Arab<< >>kur_Cyrl<< >>kur_Latn<< >>kvx<< >>kxp<< >>kyw<< >>lah<< >>lbm<< >>lhl<< >>lki<< >>lmn<< >>lrc<< >>lrl<< >>lsa<< >>lss<< >>luv<< >>luz<< >>mag<< >>mai<< >>mar<< >>mby<< >>mjl<< >>mjz<< >>mkb<< >>mke<< >>mki<< >>mnj<< >>mvy<< >>mwr<< >>mzn<< >>nag<< >>nep<< >>nhh<< >>nli<< >>nlx<< >>noe<< >>noi<< >>npi<< >>ntz<< >>nyq<< >>odk<< >>okh<< >>omr<< >>oos<< >>ori<< >>ort<< >>oru<< >>oss<< >>pal<< >>pan<< >>pan_Guru<< >>paq<< >>pcl<< >>peo<< >>pes<< >>pgg<< >>phd<< >>phl<< >>phv<< >>pli<< >>plk<< >>plp<< >>pmh<< >>prc<< >>prn<< >>prs<< >>psh<< >>psi<< >>psu<< >>pus<< >>pwr<< >>raj<< >>rat<< >>rdb<< >>rei<< >>rhg<< >>rhg_Latn<< >>rjs<< >>rkt<< >>rmi<< >>rmq<< >>rmt<< >>rmy<< >>rom<< >>rtw<< >>san<< >>san_Deva<< >>saz<< >>sbn<< >>sck<< >>scl<< >>sdb<< >>sdf<< >>sdg<< >>sdh<< >>sdr<< >>sgh<< >>sgl<< >>sgr<< >>sgy<< >>shd<< >>shm<< >>sin<< >>siy<< >>sjp<< >>skr<< >>smm<< >>smv<< >>smy<< >>snd<< >>snd_Arab<< >>sog<< >>soi<< >>soj<< >>sqo<< >>srh<< >>srx<< >>srz<< >>ssi<< >>sts<< >>syl<< >>syl_Sylo<< >>tdb<< >>tgk<< >>tgk_Cyrl<< >>tgk_Latn<< >>the<< >>thl<< >>thq<< >>thr<< >>tkb<< >>tks<< >>tkt<< >>tly<< >>tly_Latn<< >>tnv<< >>tov<< >>tra<< >>trm<< >>trw<< >>ttt<< >>urd<< >>ush<< >>vaa<< >>vaf<< >>vah<< >>vas<< >>vav<< >>ved<< >>vgr<< >>vmh<< >>wbk<< >>wbl<< >>wne<< >>wsv<< >>wtm<< >>xbc<< >>xco<< >>xka<< >>xkc<< >>xkj<< >>xkp<< >>xpr<< >>xsc<< >>xtq<< >>xvi<< >>xxx<< >>yah<< >>yai<< >>ydg<< >>zum<< >>zza<<
|
1936 |
+
- **Original Model**: [opusTCv20230926max50+bt+jhubc_transformer-big_2024-05-30.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/deu+eng+fra+por+spa-iir/opusTCv20230926max50+bt+jhubc_transformer-big_2024-05-30.zip)
|
1937 |
+
- **Resources for more information:**
|
1938 |
+
- [OPUS-MT dashboard](https://opus.nlpl.eu/dashboard/index.php?pkg=opusmt&test=all&scoreslang=all&chart=standard&model=Tatoeba-MT-models/deu%2Beng%2Bfra%2Bpor%2Bspa-iir/opusTCv20230926max50%2Bbt%2Bjhubc_transformer-big_2024-05-30)
|
1939 |
+
- [OPUS-MT-train GitHub Repo](https://github.com/Helsinki-NLP/OPUS-MT-train)
|
1940 |
+
- [More information about MarianNMT models in the transformers library](https://huggingface.co/docs/transformers/model_doc/marian)
|
1941 |
+
- [Tatoeba Translation Challenge](https://github.com/Helsinki-NLP/Tatoeba-Challenge/)
|
1942 |
+
- [HPLT bilingual data v1 (as part of the Tatoeba Translation Challenge dataset)](https://hplt-project.org/datasets/v1)
|
1943 |
+
- [A massively parallel Bible corpus](https://aclanthology.org/L14-1215/)
|
1944 |
+
|
1945 |
+
This is a multilingual translation model with multiple target languages. A sentence initial language token is required in the form of `>>id<<` (id = valid target language ID), e.g. `>>anp<<`
|
1946 |
+
|
1947 |
+
## Uses
|
1948 |
+
|
1949 |
+
This model can be used for translation and text-to-text generation.
|
1950 |
+
|
1951 |
+
## Risks, Limitations and Biases
|
1952 |
+
|
1953 |
+
**CONTENT WARNING: Readers should be aware that the model is trained on various public data sets that may contain content that is disturbing, offensive, and can propagate historical and current stereotypes.**
|
1954 |
+
|
1955 |
+
Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)).
|
1956 |
+
|
1957 |
+
## How to Get Started With the Model
|
1958 |
+
|
1959 |
+
A short example code:
|
1960 |
+
|
1961 |
+
```python
|
1962 |
+
from transformers import MarianMTModel, MarianTokenizer
|
1963 |
+
|
1964 |
+
src_text = [
|
1965 |
+
">>anp<< Replace this with text in an accepted source language.",
|
1966 |
+
">>zza<< This is the second sentence."
|
1967 |
+
]
|
1968 |
+
|
1969 |
+
model_name = "pytorch-models/opus-mt-tc-bible-big-deu_eng_fra_por_spa-iir"
|
1970 |
+
tokenizer = MarianTokenizer.from_pretrained(model_name)
|
1971 |
+
model = MarianMTModel.from_pretrained(model_name)
|
1972 |
+
translated = model.generate(**tokenizer(src_text, return_tensors="pt", padding=True))
|
1973 |
+
|
1974 |
+
for t in translated:
|
1975 |
+
print( tokenizer.decode(t, skip_special_tokens=True) )
|
1976 |
+
```
|
1977 |
+
|
1978 |
+
You can also use OPUS-MT models with the transformers pipelines, for example:
|
1979 |
+
|
1980 |
+
```python
|
1981 |
+
from transformers import pipeline
|
1982 |
+
pipe = pipeline("translation", model="Helsinki-NLP/opus-mt-tc-bible-big-deu_eng_fra_por_spa-iir")
|
1983 |
+
print(pipe(">>anp<< Replace this with text in an accepted source language."))
|
1984 |
+
```
|
1985 |
+
|
1986 |
+
## Training
|
1987 |
+
|
1988 |
+
- **Data**: opusTCv20230926max50+bt+jhubc ([source](https://github.com/Helsinki-NLP/Tatoeba-Challenge))
|
1989 |
+
- **Pre-processing**: SentencePiece (spm32k,spm32k)
|
1990 |
+
- **Model Type:** transformer-big
|
1991 |
+
- **Original MarianNMT Model**: [opusTCv20230926max50+bt+jhubc_transformer-big_2024-05-30.zip](https://object.pouta.csc.fi/Tatoeba-MT-models/deu+eng+fra+por+spa-iir/opusTCv20230926max50+bt+jhubc_transformer-big_2024-05-30.zip)
|
1992 |
+
- **Training Scripts**: [GitHub Repo](https://github.com/Helsinki-NLP/OPUS-MT-train)
|
1993 |
+
|
1994 |
+
## Evaluation
|
1995 |
+
|
1996 |
+
* [Model scores at the OPUS-MT dashboard](https://opus.nlpl.eu/dashboard/index.php?pkg=opusmt&test=all&scoreslang=all&chart=standard&model=Tatoeba-MT-models/deu%2Beng%2Bfra%2Bpor%2Bspa-iir/opusTCv20230926max50%2Bbt%2Bjhubc_transformer-big_2024-05-30)
|
1997 |
+
* test set translations: [opusTCv20230926max50+bt+jhubc_transformer-big_2024-05-29.test.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/deu+eng+fra+por+spa-iir/opusTCv20230926max50+bt+jhubc_transformer-big_2024-05-29.test.txt)
|
1998 |
+
* test set scores: [opusTCv20230926max50+bt+jhubc_transformer-big_2024-05-29.eval.txt](https://object.pouta.csc.fi/Tatoeba-MT-models/deu+eng+fra+por+spa-iir/opusTCv20230926max50+bt+jhubc_transformer-big_2024-05-29.eval.txt)
|
1999 |
+
* benchmark results: [benchmark_results.txt](benchmark_results.txt)
|
2000 |
+
* benchmark output: [benchmark_translations.zip](benchmark_translations.zip)
|
2001 |
+
|
2002 |
+
| langpair | testset | chr-F | BLEU | #sent | #words |
|
2003 |
+
|----------|---------|-------|-------|-------|--------|
|
2004 |
+
| deu-fas | tatoeba-test-v2021-08-07 | 0.45763 | 20.3 | 3185 | 24941 |
|
2005 |
+
| deu-kur_Latn | tatoeba-test-v2021-08-07 | 1.027 | 0.6 | 223 | 1249 |
|
2006 |
+
| eng-ben | tatoeba-test-v2021-08-07 | 0.47927 | 17.6 | 2500 | 11654 |
|
2007 |
+
| eng-fas | tatoeba-test-v2021-08-07 | 0.40192 | 17.1 | 3762 | 31110 |
|
2008 |
+
| eng-hin | tatoeba-test-v2021-08-07 | 0.52525 | 28.4 | 5000 | 32904 |
|
2009 |
+
| eng-kur_Latn | tatoeba-test-v2021-08-07 | 0.493 | 0.0 | 290 | 1682 |
|
2010 |
+
| eng-mar | tatoeba-test-v2021-08-07 | 0.52549 | 24.4 | 10396 | 61140 |
|
2011 |
+
| eng-pes | tatoeba-test-v2021-08-07 | 0.40401 | 17.3 | 3757 | 31044 |
|
2012 |
+
| eng-urd | tatoeba-test-v2021-08-07 | 0.45764 | 18.1 | 1663 | 12155 |
|
2013 |
+
| fra-fas | tatoeba-test-v2021-08-07 | 0.42414 | 18.9 | 376 | 3217 |
|
2014 |
+
| deu-npi | flores101-devtest | 3.082 | 0.2 | 1012 | 19762 |
|
2015 |
+
| eng-ben | flores101-devtest | 0.51055 | 17.0 | 1012 | 21155 |
|
2016 |
+
| eng-ckb | flores101-devtest | 0.45337 | 7.1 | 1012 | 21159 |
|
2017 |
+
| eng-guj | flores101-devtest | 0.53972 | 22.3 | 1012 | 23840 |
|
2018 |
+
| eng-hin | flores101-devtest | 0.57980 | 33.4 | 1012 | 27743 |
|
2019 |
+
| eng-mar | flores101-devtest | 0.48206 | 14.3 | 1012 | 21810 |
|
2020 |
+
| eng-urd | flores101-devtest | 0.48050 | 20.5 | 1012 | 28098 |
|
2021 |
+
| fra-ben | flores101-devtest | 0.43806 | 10.9 | 1012 | 21155 |
|
2022 |
+
| fra-ckb | flores101-devtest | 0.41016 | 4.9 | 1012 | 21159 |
|
2023 |
+
| por-ben | flores101-devtest | 0.42730 | 10.0 | 1012 | 21155 |
|
2024 |
+
| por-npi | flores101-devtest | 2.084 | 0.2 | 1012 | 19762 |
|
2025 |
+
| spa-hin | flores101-devtest | 0.43371 | 16.0 | 1012 | 27743 |
|
2026 |
+
| deu-ben | flores200-devtest | 0.44005 | 10.6 | 1012 | 21155 |
|
2027 |
+
| deu-hin | flores200-devtest | 0.48448 | 22.3 | 1012 | 27743 |
|
2028 |
+
| deu-hne | flores200-devtest | 0.42659 | 13.8 | 1012 | 26582 |
|
2029 |
+
| deu-mag | flores200-devtest | 0.42477 | 14.0 | 1012 | 26516 |
|
2030 |
+
| deu-npi | flores200-devtest | 5.870 | 0.1 | 1012 | 19762 |
|
2031 |
+
| deu-pes | flores200-devtest | 0.42726 | 14.9 | 1012 | 24986 |
|
2032 |
+
| deu-tgk | flores200-devtest | 0.40932 | 12.9 | 1012 | 25530 |
|
2033 |
+
| deu-urd | flores200-devtest | 0.41250 | 14.4 | 1012 | 28098 |
|
2034 |
+
| eng-ben | flores200-devtest | 0.51361 | 17.1 | 1012 | 21155 |
|
2035 |
+
| eng-ckb | flores200-devtest | 0.45750 | 7.7 | 1012 | 21152 |
|
2036 |
+
| eng-guj | flores200-devtest | 0.54231 | 22.4 | 1012 | 23840 |
|
2037 |
+
| eng-hin | flores200-devtest | 0.58371 | 33.7 | 1012 | 27743 |
|
2038 |
+
| eng-hne | flores200-devtest | 0.47591 | 19.9 | 1012 | 26582 |
|
2039 |
+
| eng-mag | flores200-devtest | 0.51070 | 22.2 | 1012 | 26516 |
|
2040 |
+
| eng-mar | flores200-devtest | 0.48733 | 14.8 | 1012 | 21810 |
|
2041 |
+
| eng-pan | flores200-devtest | 0.45015 | 18.1 | 1012 | 27451 |
|
2042 |
+
| eng-pes | flores200-devtest | 0.48588 | 21.1 | 1012 | 24986 |
|
2043 |
+
| eng-prs | flores200-devtest | 0.51879 | 24.5 | 1012 | 25885 |
|
2044 |
+
| eng-sin | flores200-devtest | 0.43823 | 10.6 | 1012 | 23278 |
|
2045 |
+
| eng-tgk | flores200-devtest | 0.47323 | 17.8 | 1012 | 25530 |
|
2046 |
+
| eng-urd | flores200-devtest | 0.48212 | 20.4 | 1012 | 28098 |
|
2047 |
+
| fra-ben | flores200-devtest | 0.44029 | 11.0 | 1012 | 21155 |
|
2048 |
+
| fra-ckb | flores200-devtest | 0.41353 | 5.3 | 1012 | 21152 |
|
2049 |
+
| fra-hin | flores200-devtest | 0.48406 | 22.6 | 1012 | 27743 |
|
2050 |
+
| fra-hne | flores200-devtest | 0.42353 | 13.9 | 1012 | 26582 |
|
2051 |
+
| fra-mag | flores200-devtest | 0.42678 | 14.3 | 1012 | 26516 |
|
2052 |
+
| fra-npi | flores200-devtest | 6.525 | 0.1 | 1012 | 19762 |
|
2053 |
+
| fra-pes | flores200-devtest | 0.43526 | 15.5 | 1012 | 24986 |
|
2054 |
+
| fra-tgk | flores200-devtest | 0.42982 | 13.7 | 1012 | 25530 |
|
2055 |
+
| fra-urd | flores200-devtest | 0.41438 | 14.2 | 1012 | 28098 |
|
2056 |
+
| por-ben | flores200-devtest | 0.43390 | 10.4 | 1012 | 21155 |
|
2057 |
+
| por-ckb | flores200-devtest | 0.42303 | 5.6 | 1012 | 21152 |
|
2058 |
+
| por-hin | flores200-devtest | 0.49524 | 23.6 | 1012 | 27743 |
|
2059 |
+
| por-hne | flores200-devtest | 0.42269 | 13.9 | 1012 | 26582 |
|
2060 |
+
| por-mag | flores200-devtest | 0.42753 | 15.0 | 1012 | 26516 |
|
2061 |
+
| por-npi | flores200-devtest | 6.737 | 0.1 | 1012 | 19762 |
|
2062 |
+
| por-pes | flores200-devtest | 0.43194 | 15.4 | 1012 | 24986 |
|
2063 |
+
| por-tgk | flores200-devtest | 0.41860 | 13.2 | 1012 | 25530 |
|
2064 |
+
| por-urd | flores200-devtest | 0.41799 | 14.8 | 1012 | 28098 |
|
2065 |
+
| spa-ben | flores200-devtest | 0.41893 | 8.3 | 1012 | 21155 |
|
2066 |
+
| spa-hin | flores200-devtest | 0.43777 | 16.4 | 1012 | 27743 |
|
2067 |
+
| spa-kas_Arab | flores200-devtest | 9.380 | 0.1 | 1012 | 23514 |
|
2068 |
+
| spa-npi | flores200-devtest | 7.518 | 0.2 | 1012 | 19762 |
|
2069 |
+
| spa-pes | flores200-devtest | 0.40856 | 12.2 | 1012 | 24986 |
|
2070 |
+
| spa-prs | flores200-devtest | 0.40361 | 12.8 | 1012 | 25885 |
|
2071 |
+
| spa-tgk | flores200-devtest | 0.40100 | 10.8 | 1012 | 25530 |
|
2072 |
+
| eng-hin | newstest2014 | 0.51249 | 23.6 | 2507 | 60872 |
|
2073 |
+
| eng-guj | newstest2019 | 0.57282 | 25.5 | 998 | 21924 |
|
2074 |
+
| deu-ben | ntrex128 | 0.43971 | 9.6 | 1997 | 40095 |
|
2075 |
+
| deu-fas | ntrex128 | 0.41469 | 13.8 | 1997 | 50525 |
|
2076 |
+
| deu-hin | ntrex128 | 0.42940 | 16.8 | 1997 | 55219 |
|
2077 |
+
| deu-snd_Arab | ntrex128 | 6.129 | 0.1 | 1997 | 49866 |
|
2078 |
+
| deu-urd | ntrex128 | 0.41881 | 14.5 | 1997 | 54259 |
|
2079 |
+
| eng-ben | ntrex128 | 0.51555 | 16.6 | 1997 | 40095 |
|
2080 |
+
| eng-fas | ntrex128 | 0.46895 | 19.7 | 1997 | 50525 |
|
2081 |
+
| eng-guj | ntrex128 | 0.48990 | 17.1 | 1997 | 45335 |
|
2082 |
+
| eng-hin | ntrex128 | 0.52307 | 26.9 | 1997 | 55219 |
|
2083 |
+
| eng-mar | ntrex128 | 0.44580 | 10.4 | 1997 | 42375 |
|
2084 |
+
| eng-nep | ntrex128 | 0.42955 | 8.4 | 1997 | 40570 |
|
2085 |
+
| eng-pan | ntrex128 | 0.46141 | 19.6 | 1997 | 54355 |
|
2086 |
+
| eng-sin | ntrex128 | 0.42236 | 9.7 | 1997 | 44429 |
|
2087 |
+
| eng-snd_Arab | ntrex128 | 1.932 | 0.1 | 1997 | 49866 |
|
2088 |
+
| eng-urd | ntrex128 | 0.49646 | 22.1 | 1997 | 54259 |
|
2089 |
+
| fra-ben | ntrex128 | 0.41716 | 8.9 | 1997 | 40095 |
|
2090 |
+
| fra-fas | ntrex128 | 0.41282 | 13.8 | 1997 | 50525 |
|
2091 |
+
| fra-hin | ntrex128 | 0.42475 | 17.1 | 1997 | 55219 |
|
2092 |
+
| fra-snd_Arab | ntrex128 | 6.047 | 0.0 | 1997 | 49866 |
|
2093 |
+
| fra-urd | ntrex128 | 0.41536 | 14.8 | 1997 | 54259 |
|
2094 |
+
| por-ben | ntrex128 | 0.43855 | 9.9 | 1997 | 40095 |
|
2095 |
+
| por-fas | ntrex128 | 0.42010 | 14.4 | 1997 | 50525 |
|
2096 |
+
| por-hin | ntrex128 | 0.43275 | 17.6 | 1997 | 55219 |
|
2097 |
+
| por-snd_Arab | ntrex128 | 6.336 | 0.1 | 1997 | 49866 |
|
2098 |
+
| por-urd | ntrex128 | 0.42484 | 15.2 | 1997 | 54259 |
|
2099 |
+
| spa-ben | ntrex128 | 0.44905 | 10.3 | 1997 | 40095 |
|
2100 |
+
| spa-fas | ntrex128 | 0.42207 | 14.1 | 1997 | 50525 |
|
2101 |
+
| spa-hin | ntrex128 | 0.43380 | 17.6 | 1997 | 55219 |
|
2102 |
+
| spa-snd_Arab | ntrex128 | 5.551 | 0.0 | 1997 | 49866 |
|
2103 |
+
| spa-urd | ntrex128 | 0.42434 | 15.0 | 1997 | 54259 |
|
2104 |
+
| eng-ben | tico19-test | 0.51563 | 17.9 | 2100 | 51695 |
|
2105 |
+
| eng-ckb | tico19-test | 0.46188 | 8.9 | 2100 | 50500 |
|
2106 |
+
| eng-fas | tico19-test | 0.53182 | 25.8 | 2100 | 59779 |
|
2107 |
+
| eng-hin | tico19-test | 0.63128 | 41.6 | 2100 | 62680 |
|
2108 |
+
| eng-mar | tico19-test | 0.45619 | 12.9 | 2100 | 50872 |
|
2109 |
+
| eng-nep | tico19-test | 0.53413 | 17.6 | 2100 | 48363 |
|
2110 |
+
| eng-prs | tico19-test | 0.44101 | 17.3 | 2100 | 62972 |
|
2111 |
+
| eng-pus | tico19-test | 0.47063 | 20.5 | 2100 | 66213 |
|
2112 |
+
| eng-urd | tico19-test | 0.51054 | 22.0 | 2100 | 65312 |
|
2113 |
+
| fra-fas | tico19-test | 0.43476 | 17.9 | 2100 | 59779 |
|
2114 |
+
| fra-hin | tico19-test | 0.48625 | 25.6 | 2100 | 62680 |
|
2115 |
+
| fra-nep | tico19-test | 0.41153 | 9.7 | 2100 | 48363 |
|
2116 |
+
| fra-urd | tico19-test | 0.40482 | 14.4 | 2100 | 65312 |
|
2117 |
+
| por-ben | tico19-test | 0.45814 | 12.5 | 2100 | 51695 |
|
2118 |
+
| por-ckb | tico19-test | 0.41684 | 5.6 | 2100 | 50500 |
|
2119 |
+
| por-fas | tico19-test | 0.49181 | 21.3 | 2100 | 59779 |
|
2120 |
+
| por-hin | tico19-test | 0.55759 | 31.1 | 2100 | 62680 |
|
2121 |
+
| por-mar | tico19-test | 0.40067 | 9.1 | 2100 | 50872 |
|
2122 |
+
| por-nep | tico19-test | 0.47378 | 12.1 | 2100 | 48363 |
|
2123 |
+
| por-pus | tico19-test | 0.42496 | 15.9 | 2100 | 66213 |
|
2124 |
+
| por-urd | tico19-test | 0.45560 | 16.6 | 2100 | 65312 |
|
2125 |
+
| spa-ben | tico19-test | 0.45751 | 12.7 | 2100 | 51695 |
|
2126 |
+
| spa-ckb | tico19-test | 0.41568 | 5.4 | 2100 | 50500 |
|
2127 |
+
| spa-fas | tico19-test | 0.48974 | 21.0 | 2100 | 59779 |
|
2128 |
+
| spa-hin | tico19-test | 0.55641 | 30.9 | 2100 | 62680 |
|
2129 |
+
| spa-mar | tico19-test | 0.40329 | 9.4 | 2100 | 50872 |
|
2130 |
+
| spa-nep | tico19-test | 0.47164 | 12.1 | 2100 | 48363 |
|
2131 |
+
| spa-prs | tico19-test | 0.41879 | 14.3 | 2100 | 62972 |
|
2132 |
+
| spa-pus | tico19-test | 0.41714 | 15.1 | 2100 | 66213 |
|
2133 |
+
| spa-urd | tico19-test | 0.44931 | 15.3 | 2100 | 65312 |
|
2134 |
+
|
2135 |
+
## Citation Information
|
2136 |
+
|
2137 |
+
* Publications: [Democratizing neural machine translation with OPUS-MT](https://doi.org/10.1007/s10579-023-09704-w) and [OPUS-MT – Building open translation services for the World](https://aclanthology.org/2020.eamt-1.61/) and [The Tatoeba Translation Challenge – Realistic Data Sets for Low Resource and Multilingual MT](https://aclanthology.org/2020.wmt-1.139/) (Please, cite if you use this model.)
|
2138 |
+
|
2139 |
+
```bibtex
|
2140 |
+
@article{tiedemann2023democratizing,
|
2141 |
+
title={Democratizing neural machine translation with {OPUS-MT}},
|
2142 |
+
author={Tiedemann, J{\"o}rg and Aulamo, Mikko and Bakshandaeva, Daria and Boggia, Michele and Gr{\"o}nroos, Stig-Arne and Nieminen, Tommi and Raganato, Alessandro and Scherrer, Yves and Vazquez, Raul and Virpioja, Sami},
|
2143 |
+
journal={Language Resources and Evaluation},
|
2144 |
+
number={58},
|
2145 |
+
pages={713--755},
|
2146 |
+
year={2023},
|
2147 |
+
publisher={Springer Nature},
|
2148 |
+
issn={1574-0218},
|
2149 |
+
doi={10.1007/s10579-023-09704-w}
|
2150 |
+
}
|
2151 |
+
|
2152 |
+
@inproceedings{tiedemann-thottingal-2020-opus,
|
2153 |
+
title = "{OPUS}-{MT} {--} Building open translation services for the World",
|
2154 |
+
author = {Tiedemann, J{\"o}rg and Thottingal, Santhosh},
|
2155 |
+
booktitle = "Proceedings of the 22nd Annual Conference of the European Association for Machine Translation",
|
2156 |
+
month = nov,
|
2157 |
+
year = "2020",
|
2158 |
+
address = "Lisboa, Portugal",
|
2159 |
+
publisher = "European Association for Machine Translation",
|
2160 |
+
url = "https://aclanthology.org/2020.eamt-1.61",
|
2161 |
+
pages = "479--480",
|
2162 |
+
}
|
2163 |
+
|
2164 |
+
@inproceedings{tiedemann-2020-tatoeba,
|
2165 |
+
title = "The Tatoeba Translation Challenge {--} Realistic Data Sets for Low Resource and Multilingual {MT}",
|
2166 |
+
author = {Tiedemann, J{\"o}rg},
|
2167 |
+
booktitle = "Proceedings of the Fifth Conference on Machine Translation",
|
2168 |
+
month = nov,
|
2169 |
+
year = "2020",
|
2170 |
+
address = "Online",
|
2171 |
+
publisher = "Association for Computational Linguistics",
|
2172 |
+
url = "https://aclanthology.org/2020.wmt-1.139",
|
2173 |
+
pages = "1174--1182",
|
2174 |
+
}
|
2175 |
+
```
|
2176 |
+
|
2177 |
+
## Acknowledgements
|
2178 |
+
|
2179 |
+
The work is supported by the [HPLT project](https://hplt-project.org/), funded by the European Union’s Horizon Europe research and innovation programme under grant agreement No 101070350. We are also grateful for the generous computational resources and IT infrastructure provided by [CSC -- IT Center for Science](https://www.csc.fi/), Finland, and the [EuroHPC supercomputer LUMI](https://www.lumi-supercomputer.eu/).
|
2180 |
+
|
2181 |
+
## Model conversion info
|
2182 |
+
|
2183 |
+
* transformers version: 4.45.1
|
2184 |
+
* OPUS-MT git hash: 0882077
|
2185 |
+
* port time: Tue Oct 8 10:05:20 EEST 2024
|
2186 |
+
* port machine: LM0-400-22516.local
|
benchmark_results.txt
ADDED
@@ -0,0 +1,263 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
multi-multi tatoeba-test-v2020-07-28-v2023-09-26 0.45945 19.5 10000 65277
|
2 |
+
deu-asm flores101-devtest 0.23909 2.6 1012 21028
|
3 |
+
deu-ckb flores101-devtest 0.39919 4.3 1012 21159
|
4 |
+
deu-npi flores101-devtest 3.082 0.2 1012 19762
|
5 |
+
deu-pan flores101-devtest 0.36883 10.9 1012 27451
|
6 |
+
eng-ben flores101-devtest 0.51055 17.0 1012 21155
|
7 |
+
eng-ckb flores101-devtest 0.45337 7.1 1012 21159
|
8 |
+
eng-guj flores101-devtest 0.53972 22.3 1012 23840
|
9 |
+
eng-hin flores101-devtest 0.57980 33.4 1012 27743
|
10 |
+
eng-mar flores101-devtest 0.48206 14.3 1012 21810
|
11 |
+
eng-pus flores101-devtest 0.37264 11.9 1012 27953
|
12 |
+
eng-urd flores101-devtest 0.48050 20.5 1012 28098
|
13 |
+
fra-asm flores101-devtest 0.24378 2.4 1012 21028
|
14 |
+
fra-ben flores101-devtest 0.43806 10.9 1012 21155
|
15 |
+
fra-ckb flores101-devtest 0.41016 4.9 1012 21159
|
16 |
+
fra-pan flores101-devtest 0.37066 11.0 1012 27451
|
17 |
+
fra-pus flores101-devtest 0.33511 8.5 1012 27953
|
18 |
+
por-ben flores101-devtest 0.42730 10.0 1012 21155
|
19 |
+
por-npi flores101-devtest 2.084 0.2 1012 19762
|
20 |
+
por-pan flores101-devtest 0.36551 10.7 1012 27451
|
21 |
+
spa-ckb flores101-devtest 0.39331 3.9 1012 21159
|
22 |
+
spa-hin flores101-devtest 0.43371 16.0 1012 27743
|
23 |
+
spa-tgk flores101-devtest 0.39762 10.6 1012 25530
|
24 |
+
deu-asm flores200-devtest 0.24326 2.8 1012 21028
|
25 |
+
deu-awa flores200-devtest 0.26471 4.9 1012 26642
|
26 |
+
deu-ben flores200-devtest 0.44005 10.6 1012 21155
|
27 |
+
deu-bho flores200-devtest 0.34979 9.2 1012 27914
|
28 |
+
deu-ckb flores200-devtest 0.39987 4.3 1012 21152
|
29 |
+
deu-guj flores200-devtest 0.39522 11.1 1012 23840
|
30 |
+
deu-hin flores200-devtest 0.48448 22.3 1012 27743
|
31 |
+
deu-hne flores200-devtest 0.42659 13.8 1012 26582
|
32 |
+
deu-kas_Arab flores200-devtest 0.11521 0.3 1012 23514
|
33 |
+
deu-kas_Deva flores200-devtest 0.14151 0.8 1012 26371
|
34 |
+
deu-kmr flores200-devtest 0.27796 4.2 1012 25581
|
35 |
+
deu-mag flores200-devtest 0.42477 14.0 1012 26516
|
36 |
+
deu-mai flores200-devtest 0.39256 8.5 1012 25999
|
37 |
+
deu-mar flores200-devtest 0.39453 9.1 1012 21810
|
38 |
+
deu-npi flores200-devtest 5.870 0.1 1012 19762
|
39 |
+
deu-pan flores200-devtest 0.37308 11.3 1012 27451
|
40 |
+
deu-pes flores200-devtest 0.42726 14.9 1012 24986
|
41 |
+
deu-prs flores200-devtest 0.38830 12.9 1012 25885
|
42 |
+
deu-san flores200-devtest 0.15747 0.6 1012 18253
|
43 |
+
deu-sin flores200-devtest 0.37614 7.2 1012 23278
|
44 |
+
deu-tgk flores200-devtest 0.40932 12.9 1012 25530
|
45 |
+
deu-urd flores200-devtest 0.41250 14.4 1012 28098
|
46 |
+
eng-asm flores200-devtest 0.35879 5.7 1012 21028
|
47 |
+
eng-awa flores200-devtest 0.16193 2.2 1012 26642
|
48 |
+
eng-ben flores200-devtest 0.51361 17.1 1012 21155
|
49 |
+
eng-bho flores200-devtest 0.38188 12.1 1012 27914
|
50 |
+
eng-ckb flores200-devtest 0.45750 7.7 1012 21152
|
51 |
+
eng-guj flores200-devtest 0.54231 22.4 1012 23840
|
52 |
+
eng-hin flores200-devtest 0.58371 33.7 1012 27743
|
53 |
+
eng-hne flores200-devtest 0.47591 19.9 1012 26582
|
54 |
+
eng-kas_Arab flores200-devtest 0.10418 0.3 1012 23514
|
55 |
+
eng-kas_Deva flores200-devtest 0.13268 1.4 1012 26371
|
56 |
+
eng-kmr flores200-devtest 0.27424 4.8 1012 25581
|
57 |
+
eng-mag flores200-devtest 0.51070 22.2 1012 26516
|
58 |
+
eng-mai flores200-devtest 0.39249 10.0 1012 25999
|
59 |
+
eng-mar flores200-devtest 0.48733 14.8 1012 21810
|
60 |
+
eng-npi flores200-devtest 0.18525 0.4 1012 19762
|
61 |
+
eng-pan flores200-devtest 0.45015 18.1 1012 27451
|
62 |
+
eng-pes flores200-devtest 0.48588 21.1 1012 24986
|
63 |
+
eng-prs flores200-devtest 0.51879 24.5 1012 25885
|
64 |
+
eng-san flores200-devtest 0.14791 0.8 1012 18253
|
65 |
+
eng-sin flores200-devtest 0.43823 10.6 1012 23278
|
66 |
+
eng-tgk flores200-devtest 0.47323 17.8 1012 25530
|
67 |
+
eng-urd flores200-devtest 0.48212 20.4 1012 28098
|
68 |
+
fra-asm flores200-devtest 0.24451 2.7 1012 21028
|
69 |
+
fra-awa flores200-devtest 0.26762 5.0 1012 26642
|
70 |
+
fra-ben flores200-devtest 0.44029 11.0 1012 21155
|
71 |
+
fra-bho flores200-devtest 0.34314 8.9 1012 27914
|
72 |
+
fra-ckb flores200-devtest 0.41353 5.3 1012 21152
|
73 |
+
fra-guj flores200-devtest 0.38634 11.0 1012 23840
|
74 |
+
fra-hin flores200-devtest 0.48406 22.6 1012 27743
|
75 |
+
fra-hne flores200-devtest 0.42353 13.9 1012 26582
|
76 |
+
fra-kas_Arab flores200-devtest 0.12483 0.3 1012 23514
|
77 |
+
fra-kas_Deva flores200-devtest 0.14979 0.7 1012 26371
|
78 |
+
fra-kmr flores200-devtest 0.28891 4.9 1012 25581
|
79 |
+
fra-mag flores200-devtest 0.42678 14.3 1012 26516
|
80 |
+
fra-mai flores200-devtest 0.39474 8.7 1012 25999
|
81 |
+
fra-mar flores200-devtest 0.39137 8.9 1012 21810
|
82 |
+
fra-npi flores200-devtest 6.525 0.1 1012 19762
|
83 |
+
fra-pan flores200-devtest 0.36980 10.6 1012 27451
|
84 |
+
fra-pes flores200-devtest 0.43526 15.5 1012 24986
|
85 |
+
fra-prs flores200-devtest 0.37985 12.8 1012 25885
|
86 |
+
fra-san flores200-devtest 0.15075 0.6 1012 18253
|
87 |
+
fra-sin flores200-devtest 0.38137 7.2 1012 23278
|
88 |
+
fra-tgk flores200-devtest 0.42982 13.7 1012 25530
|
89 |
+
fra-urd flores200-devtest 0.41438 14.2 1012 28098
|
90 |
+
por-asm flores200-devtest 0.25163 3.0 1012 21028
|
91 |
+
por-awa flores200-devtest 0.24010 4.2 1012 26642
|
92 |
+
por-ben flores200-devtest 0.43390 10.4 1012 21155
|
93 |
+
por-bho flores200-devtest 0.34843 9.2 1012 27914
|
94 |
+
por-ckb flores200-devtest 0.42303 5.6 1012 21152
|
95 |
+
por-guj flores200-devtest 0.37374 10.6 1012 23840
|
96 |
+
por-hin flores200-devtest 0.49524 23.6 1012 27743
|
97 |
+
por-hne flores200-devtest 0.42269 13.9 1012 26582
|
98 |
+
por-kas_Arab flores200-devtest 0.11212 0.2 1012 23514
|
99 |
+
por-kas_Deva flores200-devtest 0.14385 0.9 1012 26371
|
100 |
+
por-kmr flores200-devtest 0.27537 4.3 1012 25581
|
101 |
+
por-mag flores200-devtest 0.42753 15.0 1012 26516
|
102 |
+
por-mai flores200-devtest 0.39246 8.5 1012 25999
|
103 |
+
por-mar flores200-devtest 0.38513 8.8 1012 21810
|
104 |
+
por-npi flores200-devtest 6.737 0.1 1012 19762
|
105 |
+
por-pan flores200-devtest 0.36653 10.7 1012 27451
|
106 |
+
por-pes flores200-devtest 0.43194 15.4 1012 24986
|
107 |
+
por-prs flores200-devtest 0.36411 12.1 1012 25885
|
108 |
+
por-san flores200-devtest 0.15025 0.6 1012 18253
|
109 |
+
por-sin flores200-devtest 0.38166 7.5 1012 23278
|
110 |
+
por-tgk flores200-devtest 0.41860 13.2 1012 25530
|
111 |
+
por-urd flores200-devtest 0.41799 14.8 1012 28098
|
112 |
+
spa-asm flores200-devtest 0.24079 2.3 1012 21028
|
113 |
+
spa-awa flores200-devtest 0.23696 3.7 1012 26642
|
114 |
+
spa-ben flores200-devtest 0.41893 8.3 1012 21155
|
115 |
+
spa-bho flores200-devtest 0.33382 7.9 1012 27914
|
116 |
+
spa-ckb flores200-devtest 0.39324 3.9 1012 21152
|
117 |
+
spa-guj flores200-devtest 0.37256 8.1 1012 23840
|
118 |
+
spa-hin flores200-devtest 0.43777 16.4 1012 27743
|
119 |
+
spa-hne flores200-devtest 0.39492 11.2 1012 26582
|
120 |
+
spa-kas_Arab flores200-devtest 9.380 0.1 1012 23514
|
121 |
+
spa-kas_Deva flores200-devtest 0.14350 0.7 1012 26371
|
122 |
+
spa-kmr flores200-devtest 0.26820 3.7 1012 25581
|
123 |
+
spa-mag flores200-devtest 0.39882 11.4 1012 26516
|
124 |
+
spa-mai flores200-devtest 0.36804 6.5 1012 25999
|
125 |
+
spa-mar flores200-devtest 0.35238 6.2 1012 21810
|
126 |
+
spa-npi flores200-devtest 7.518 0.2 1012 19762
|
127 |
+
spa-pan flores200-devtest 0.34618 8.4 1012 27451
|
128 |
+
spa-pes flores200-devtest 0.40856 12.2 1012 24986
|
129 |
+
spa-prs flores200-devtest 0.40361 12.8 1012 25885
|
130 |
+
spa-san flores200-devtest 0.14398 0.4 1012 18253
|
131 |
+
spa-sin flores200-devtest 0.35568 5.5 1012 23278
|
132 |
+
spa-tgk flores200-devtest 0.40100 10.8 1012 25530
|
133 |
+
spa-urd flores200-devtest 0.38539 10.9 1012 28098
|
134 |
+
eng-hin newstest2014 0.51249 23.6 2507 60872
|
135 |
+
eng-guj newstest2019 0.57282 25.5 998 21924
|
136 |
+
eng-pus newstest2020 0.32214 8.3 2719 58339
|
137 |
+
deu-ben ntrex128 0.43971 9.6 1997 40095
|
138 |
+
deu-div ntrex128 0.17330 0.2 1997 37802
|
139 |
+
deu-fas ntrex128 0.41469 13.8 1997 50525
|
140 |
+
deu-guj ntrex128 0.37353 8.4 1997 45335
|
141 |
+
deu-hin ntrex128 0.42940 16.8 1997 55219
|
142 |
+
deu-kmr ntrex128 0.28341 4.3 1997 47989
|
143 |
+
deu-mar ntrex128 0.36169 6.2 1997 42375
|
144 |
+
deu-nep ntrex128 0.36701 5.4 1997 40570
|
145 |
+
deu-pan ntrex128 0.36776 11.0 1997 54355
|
146 |
+
deu-prs ntrex128 0.34876 8.9 1997 52316
|
147 |
+
deu-pus ntrex128 0.31900 7.6 1997 57486
|
148 |
+
deu-sin ntrex128 0.36244 6.0 1997 44429
|
149 |
+
deu-snd_Arab ntrex128 6.129 0.1 1997 49866
|
150 |
+
deu-tgk_Cyrl ntrex128 0.36407 9.4 1997 48894
|
151 |
+
deu-urd ntrex128 0.41881 14.5 1997 54259
|
152 |
+
eng-ben ntrex128 0.51555 16.6 1997 40095
|
153 |
+
eng-div ntrex128 0.16820 0.2 1997 37802
|
154 |
+
eng-fas ntrex128 0.46895 19.7 1997 50525
|
155 |
+
eng-guj ntrex128 0.48990 17.1 1997 45335
|
156 |
+
eng-hin ntrex128 0.52307 26.9 1997 55219
|
157 |
+
eng-kmr ntrex128 0.26453 4.7 1997 47989
|
158 |
+
eng-mar ntrex128 0.44580 10.4 1997 42375
|
159 |
+
eng-nep ntrex128 0.42955 8.4 1997 40570
|
160 |
+
eng-pan ntrex128 0.46141 19.6 1997 54355
|
161 |
+
eng-prs ntrex128 0.39651 12.9 1997 52316
|
162 |
+
eng-pus ntrex128 0.33816 8.9 1997 57486
|
163 |
+
eng-sin ntrex128 0.42236 9.7 1997 44429
|
164 |
+
eng-snd_Arab ntrex128 1.932 0.1 1997 49866
|
165 |
+
eng-tgk_Cyrl ntrex128 0.38524 11.3 1997 48894
|
166 |
+
eng-urd ntrex128 0.49646 22.1 1997 54259
|
167 |
+
fra-ben ntrex128 0.41716 8.9 1997 40095
|
168 |
+
fra-div ntrex128 0.17398 0.2 1997 37802
|
169 |
+
fra-fas ntrex128 0.41282 13.8 1997 50525
|
170 |
+
fra-guj ntrex128 0.36516 8.0 1997 45335
|
171 |
+
fra-hin ntrex128 0.42475 17.1 1997 55219
|
172 |
+
fra-kmr ntrex128 0.28136 4.4 1997 47989
|
173 |
+
fra-mar ntrex128 0.36150 6.6 1997 42375
|
174 |
+
fra-nep ntrex128 0.35986 5.4 1997 40570
|
175 |
+
fra-pan ntrex128 0.36120 10.2 1997 54355
|
176 |
+
fra-prs ntrex128 0.33913 8.3 1997 52316
|
177 |
+
fra-pus ntrex128 0.31411 7.2 1997 57486
|
178 |
+
fra-sin ntrex128 0.36147 6.1 1997 44429
|
179 |
+
fra-snd_Arab ntrex128 6.047 0.0 1997 49866
|
180 |
+
fra-tgk_Cyrl ntrex128 0.36439 8.9 1997 48894
|
181 |
+
fra-urd ntrex128 0.41536 14.8 1997 54259
|
182 |
+
por-ben ntrex128 0.43855 9.9 1997 40095
|
183 |
+
por-div ntrex128 0.17701 0.2 1997 37802
|
184 |
+
por-fas ntrex128 0.42010 14.4 1997 50525
|
185 |
+
por-guj ntrex128 0.35961 7.8 1997 45335
|
186 |
+
por-hin ntrex128 0.43275 17.6 1997 55219
|
187 |
+
por-kmr ntrex128 0.28936 4.6 1997 47989
|
188 |
+
por-mar ntrex128 0.35835 6.6 1997 42375
|
189 |
+
por-nep ntrex128 0.36993 5.7 1997 40570
|
190 |
+
por-pan ntrex128 0.36360 10.6 1997 54355
|
191 |
+
por-prs ntrex128 0.32716 8.0 1997 52316
|
192 |
+
por-pus ntrex128 0.32074 7.7 1997 57486
|
193 |
+
por-sin ntrex128 0.36553 6.1 1997 44429
|
194 |
+
por-snd_Arab ntrex128 6.336 0.1 1997 49866
|
195 |
+
por-tgk_Cyrl ntrex128 0.37294 9.7 1997 48894
|
196 |
+
por-urd ntrex128 0.42484 15.2 1997 54259
|
197 |
+
spa-ben ntrex128 0.44905 10.3 1997 40095
|
198 |
+
spa-div ntrex128 0.17655 0.2 1997 37802
|
199 |
+
spa-fas ntrex128 0.42207 14.1 1997 50525
|
200 |
+
spa-guj ntrex128 0.38435 8.5 1997 45335
|
201 |
+
spa-hin ntrex128 0.43380 17.6 1997 55219
|
202 |
+
spa-kmr ntrex128 0.28968 4.5 1997 47989
|
203 |
+
spa-mar ntrex128 0.36052 6.3 1997 42375
|
204 |
+
spa-nep ntrex128 0.37196 5.5 1997 40570
|
205 |
+
spa-pan ntrex128 0.37361 11.1 1997 54355
|
206 |
+
spa-prs ntrex128 0.37448 10.3 1997 52316
|
207 |
+
spa-pus ntrex128 0.32179 7.6 1997 57486
|
208 |
+
spa-sin ntrex128 0.36971 6.5 1997 44429
|
209 |
+
spa-snd_Arab ntrex128 5.551 0.0 1997 49866
|
210 |
+
spa-tgk_Cyrl ntrex128 0.37469 9.5 1997 48894
|
211 |
+
spa-urd ntrex128 0.42434 15.0 1997 54259
|
212 |
+
eng-awa tatoeba-test-v2021-03-30 0.19164 3.0 280 1151
|
213 |
+
eng-pes tatoeba-test-v2021-03-30 0.40069 17.1 3763 31069
|
214 |
+
eng-zza tatoeba-test-v2021-03-30 0.11257 1.1 533 3038
|
215 |
+
deu-fas tatoeba-test-v2021-08-07 0.45763 20.3 3185 24941
|
216 |
+
deu-kur_Latn tatoeba-test-v2021-08-07 1.027 0.6 223 1249
|
217 |
+
eng-awa tatoeba-test-v2021-08-07 0.18842 3.0 279 1148
|
218 |
+
eng-ben tatoeba-test-v2021-08-07 0.47927 17.6 2500 11654
|
219 |
+
eng-fas tatoeba-test-v2021-08-07 0.40192 17.1 3762 31110
|
220 |
+
eng-hin tatoeba-test-v2021-08-07 0.52525 28.4 5000 32904
|
221 |
+
eng-kur_Latn tatoeba-test-v2021-08-07 0.493 0.0 290 1682
|
222 |
+
eng-mar tatoeba-test-v2021-08-07 0.52549 24.4 10396 61140
|
223 |
+
eng-pes tatoeba-test-v2021-08-07 0.40401 17.3 3757 31044
|
224 |
+
eng-rom tatoeba-test-v2021-08-07 0.21680 1.6 706 5222
|
225 |
+
eng-urd tatoeba-test-v2021-08-07 0.45764 18.1 1663 12155
|
226 |
+
eng-zza tatoeba-test-v2021-08-07 0.10823 0.8 529 3015
|
227 |
+
fra-fas tatoeba-test-v2021-08-07 0.42414 18.9 376 3217
|
228 |
+
eng-ben tico19-test 0.51563 17.9 2100 51695
|
229 |
+
eng-ckb tico19-test 0.46188 8.9 2100 50500
|
230 |
+
eng-fas tico19-test 0.53182 25.8 2100 59779
|
231 |
+
eng-hin tico19-test 0.63128 41.6 2100 62680
|
232 |
+
eng-mar tico19-test 0.45619 12.9 2100 50872
|
233 |
+
eng-nep tico19-test 0.53413 17.6 2100 48363
|
234 |
+
eng-prs tico19-test 0.44101 17.3 2100 62972
|
235 |
+
eng-pus tico19-test 0.47063 20.5 2100 66213
|
236 |
+
eng-urd tico19-test 0.51054 22.0 2100 65312
|
237 |
+
fra-ben tico19-test 0.39422 9.6 2100 51695
|
238 |
+
fra-ckb tico19-test 0.37776 4.9 2100 50500
|
239 |
+
fra-fas tico19-test 0.43476 17.9 2100 59779
|
240 |
+
fra-hin tico19-test 0.48625 25.6 2100 62680
|
241 |
+
fra-mar tico19-test 0.36498 7.7 2100 50872
|
242 |
+
fra-nep tico19-test 0.41153 9.7 2100 48363
|
243 |
+
fra-prs tico19-test 0.36130 11.6 2100 62972
|
244 |
+
fra-pus tico19-test 0.37217 12.7 2100 66213
|
245 |
+
fra-urd tico19-test 0.40482 14.4 2100 65312
|
246 |
+
por-ben tico19-test 0.45814 12.5 2100 51695
|
247 |
+
por-ckb tico19-test 0.41684 5.6 2100 50500
|
248 |
+
por-fas tico19-test 0.49181 21.3 2100 59779
|
249 |
+
por-hin tico19-test 0.55759 31.1 2100 62680
|
250 |
+
por-mar tico19-test 0.40067 9.1 2100 50872
|
251 |
+
por-nep tico19-test 0.47378 12.1 2100 48363
|
252 |
+
por-prs tico19-test 0.38725 12.1 2100 62972
|
253 |
+
por-pus tico19-test 0.42496 15.9 2100 66213
|
254 |
+
por-urd tico19-test 0.45560 16.6 2100 65312
|
255 |
+
spa-ben tico19-test 0.45751 12.7 2100 51695
|
256 |
+
spa-ckb tico19-test 0.41568 5.4 2100 50500
|
257 |
+
spa-fas tico19-test 0.48974 21.0 2100 59779
|
258 |
+
spa-hin tico19-test 0.55641 30.9 2100 62680
|
259 |
+
spa-mar tico19-test 0.40329 9.4 2100 50872
|
260 |
+
spa-nep tico19-test 0.47164 12.1 2100 48363
|
261 |
+
spa-prs tico19-test 0.41879 14.3 2100 62972
|
262 |
+
spa-pus tico19-test 0.41714 15.1 2100 66213
|
263 |
+
spa-urd tico19-test 0.44931 15.3 2100 65312
|
benchmark_translations.zip
ADDED
File without changes
|
config.json
ADDED
@@ -0,0 +1,41 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "pytorch-models/opus-mt-tc-bible-big-deu_eng_fra_por_spa-iir",
|
3 |
+
"activation_dropout": 0.0,
|
4 |
+
"activation_function": "relu",
|
5 |
+
"architectures": [
|
6 |
+
"MarianMTModel"
|
7 |
+
],
|
8 |
+
"attention_dropout": 0.0,
|
9 |
+
"bos_token_id": 0,
|
10 |
+
"classifier_dropout": 0.0,
|
11 |
+
"d_model": 1024,
|
12 |
+
"decoder_attention_heads": 16,
|
13 |
+
"decoder_ffn_dim": 4096,
|
14 |
+
"decoder_layerdrop": 0.0,
|
15 |
+
"decoder_layers": 6,
|
16 |
+
"decoder_start_token_id": 62089,
|
17 |
+
"decoder_vocab_size": 62090,
|
18 |
+
"dropout": 0.1,
|
19 |
+
"encoder_attention_heads": 16,
|
20 |
+
"encoder_ffn_dim": 4096,
|
21 |
+
"encoder_layerdrop": 0.0,
|
22 |
+
"encoder_layers": 6,
|
23 |
+
"eos_token_id": 467,
|
24 |
+
"forced_eos_token_id": null,
|
25 |
+
"init_std": 0.02,
|
26 |
+
"is_encoder_decoder": true,
|
27 |
+
"max_length": null,
|
28 |
+
"max_position_embeddings": 1024,
|
29 |
+
"model_type": "marian",
|
30 |
+
"normalize_embedding": false,
|
31 |
+
"num_beams": null,
|
32 |
+
"num_hidden_layers": 6,
|
33 |
+
"pad_token_id": 62089,
|
34 |
+
"scale_embedding": true,
|
35 |
+
"share_encoder_decoder_embeddings": true,
|
36 |
+
"static_position_embeddings": true,
|
37 |
+
"torch_dtype": "float32",
|
38 |
+
"transformers_version": "4.45.1",
|
39 |
+
"use_cache": true,
|
40 |
+
"vocab_size": 62090
|
41 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bad_words_ids": [
|
4 |
+
[
|
5 |
+
62089
|
6 |
+
]
|
7 |
+
],
|
8 |
+
"bos_token_id": 0,
|
9 |
+
"decoder_start_token_id": 62089,
|
10 |
+
"eos_token_id": 467,
|
11 |
+
"forced_eos_token_id": 467,
|
12 |
+
"max_length": 512,
|
13 |
+
"num_beams": 4,
|
14 |
+
"pad_token_id": 62089,
|
15 |
+
"transformers_version": "4.45.1"
|
16 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:90e061992617affc8a12ead47556375bf923d5ef06c5982fc7e05d93ed2cf65c
|
3 |
+
size 960028120
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d07aae1331416f131d30a2b561547fbfef33711dc053aafe4e17acbd84549ad4
|
3 |
+
size 960079365
|
source.spm
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:53c59218af2fc5810fc2016b81e59f0292761116929ace43d998350513d874f3
|
3 |
+
size 802230
|
special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"}
|
target.spm
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:40c5f34005866ed800b90815e0c220112e5e3e6044eedaf37d9fa03b0e77a21b
|
3 |
+
size 924635
|
tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"source_lang": "deu+eng+fra+por+spa", "target_lang": "iir", "unk_token": "<unk>", "eos_token": "</s>", "pad_token": "<pad>", "model_max_length": 512, "sp_model_kwargs": {}, "separate_vocabs": false, "special_tokens_map_file": null, "name_or_path": "marian-models/opusTCv20230926max50+bt+jhubc_transformer-big_2024-05-30/deu+eng+fra+por+spa-iir", "tokenizer_class": "MarianTokenizer"}
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|