Locutusque
commited on
Commit
•
a954f40
1
Parent(s):
8f396cb
Update README.md
Browse files
README.md
CHANGED
@@ -33,7 +33,7 @@ The model architecture used in this model is GPT-2, a transformer-based language
|
|
33 |
The model is evaluated based on several metrics, including loss, reward, penalty, BLEU score, and perplexity. The loss metric is calculated during training and reflects the difference between the predicted output and the actual output. The reward metric is based on the number of correct words generated by the model, while the penalty metric penalizes the model for repeating words consecutively. The BLEU score measures the similarity between the generated text and the ground truth text, while the perplexity metric measures how well the model is able to predict the next word in a sequence. During validation, the model achieved the following metrics:
|
34 |
|
35 |
- BLEU score: 52
|
36 |
-
- Accuracy: 53
|
37 |
- perplexity: 4.3
|
38 |
|
39 |
## Limitations and Bias
|
|
|
33 |
The model is evaluated based on several metrics, including loss, reward, penalty, BLEU score, and perplexity. The loss metric is calculated during training and reflects the difference between the predicted output and the actual output. The reward metric is based on the number of correct words generated by the model, while the penalty metric penalizes the model for repeating words consecutively. The BLEU score measures the similarity between the generated text and the ground truth text, while the perplexity metric measures how well the model is able to predict the next word in a sequence. During validation, the model achieved the following metrics:
|
34 |
|
35 |
- BLEU score: 52
|
36 |
+
- Accuracy: 53
|
37 |
- perplexity: 4.3
|
38 |
|
39 |
## Limitations and Bias
|