BounharAbdelaziz
commited on
Commit
•
711e129
1
Parent(s):
1ca7c48
Update README.md
Browse files
README.md
CHANGED
@@ -23,23 +23,9 @@ It achieves the following results on the evaluation set:
|
|
23 |
- Bleu: 8.3292
|
24 |
- Gen Len: 34.4959
|
25 |
|
26 |
-
The finetuning was conducted using a
|
27 |
-
|
28 |
-
|
29 |
-
## Training hyperparameters
|
30 |
-
|
31 |
-
The following hyperparameters were used during training:
|
32 |
-
- learning_rate: 3e-05
|
33 |
-
- train_batch_size: 22
|
34 |
-
- eval_batch_size: 22
|
35 |
-
- seed: 42
|
36 |
-
- gradient_accumulation_steps: 4
|
37 |
-
- total_train_batch_size: 88
|
38 |
-
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
39 |
-
- lr_scheduler_type: linear
|
40 |
-
- lr_scheduler_warmup_ratio: 0.03
|
41 |
-
- num_epochs: 40
|
42 |
|
|
|
43 |
|
44 |
## Usage
|
45 |
|
@@ -73,9 +59,9 @@ print("Translation:", output_text)
|
|
73 |
|
74 |
Let's see an example of transliterating Moroccan Darija Arabizi to Arabic:
|
75 |
|
76 |
-
**Input**: "
|
77 |
|
78 |
-
**Output**: "مرحبا
|
79 |
|
80 |
## Limiations
|
81 |
|
@@ -88,6 +74,20 @@ We're continuously striving to improve our model's performance and usability and
|
|
88 |
If you have any feedback, suggestions, or encounter any issues, please don't hesitate to reach out to us.
|
89 |
|
90 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
91 |
## Training results
|
92 |
|
93 |
| Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
|
|
|
23 |
- Bleu: 8.3292
|
24 |
- Gen Len: 34.4959
|
25 |
|
26 |
+
The finetuning was conducted using a **A100-40GB** and took **23 hours**.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
27 |
|
28 |
+
Try it out on our dedicated [Terjman-Large Space](https://huggingface.co/spaces/atlasia/Terjman-Large) 🤗
|
29 |
|
30 |
## Usage
|
31 |
|
|
|
59 |
|
60 |
Let's see an example of transliterating Moroccan Darija Arabizi to Arabic:
|
61 |
|
62 |
+
**Input**: "Hi my friend, can you tell me a joke in moroccan darija? I'd be happy to hear that from you!"
|
63 |
|
64 |
+
**Output**: "مرحبا صديقي، يمكن لك تقول لي نكتة في داريجا المغربية؟ سأكون سعيدا بسماعها منك!"
|
65 |
|
66 |
## Limiations
|
67 |
|
|
|
74 |
If you have any feedback, suggestions, or encounter any issues, please don't hesitate to reach out to us.
|
75 |
|
76 |
|
77 |
+
## Training hyperparameters
|
78 |
+
|
79 |
+
The following hyperparameters were used during training:
|
80 |
+
- learning_rate: 3e-05
|
81 |
+
- train_batch_size: 22
|
82 |
+
- eval_batch_size: 22
|
83 |
+
- seed: 42
|
84 |
+
- gradient_accumulation_steps: 4
|
85 |
+
- total_train_batch_size: 88
|
86 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
87 |
+
- lr_scheduler_type: linear
|
88 |
+
- lr_scheduler_warmup_ratio: 0.03
|
89 |
+
- num_epochs: 40
|
90 |
+
|
91 |
## Training results
|
92 |
|
93 |
| Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
|