File size: 2,633 Bytes
9b6b578 7320b69 9b6b578 7320b69 9b6b578 9bcdbb5 bd725a7 4ed6107 cc8f52c 2117a6d 182964c f0e6393 dee7e75 959b130 ed993f1 4b1283c 11ea75c 9d023ba bb2948f 0ceb3ed e45051a 93fdbea 0dd7b6b 3beff54 aa9992f 9b5a2c7 c8bded5 7320b69 9b6b578 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 |
---
license: mit
base_model: badokorach/afriqa_afroxlmr_squad_v2-luganda
tags:
- generated_from_keras_callback
model-index:
- name: badokorach/afriqa_afroxlmr_squad_v2-luganda_311223
results: []
---
<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->
# badokorach/afriqa_afroxlmr_squad_v2-luganda_311223
This model is a fine-tuned version of [badokorach/afriqa_afroxlmr_squad_v2-luganda](https://huggingface.co/badokorach/afriqa_afroxlmr_squad_v2-luganda) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.6478
- Validation Loss: 0.0
- Epoch: 23
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 1e-05, 'decay_steps': 14760, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.02}
- training_precision: mixed_float16
### Training results
| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 3.4699 | 0.0 | 0 |
| 3.3401 | 0.0 | 1 |
| 3.1487 | 0.0 | 2 |
| 2.9589 | 0.0 | 3 |
| 2.7472 | 0.0 | 4 |
| 2.5135 | 0.0 | 5 |
| 2.3000 | 0.0 | 6 |
| 2.0904 | 0.0 | 7 |
| 1.9314 | 0.0 | 8 |
| 1.7617 | 0.0 | 9 |
| 1.6075 | 0.0 | 10 |
| 1.4573 | 0.0 | 11 |
| 1.3275 | 0.0 | 12 |
| 1.2261 | 0.0 | 13 |
| 1.1378 | 0.0 | 14 |
| 1.0358 | 0.0 | 15 |
| 0.9618 | 0.0 | 16 |
| 0.9082 | 0.0 | 17 |
| 0.8762 | 0.0 | 18 |
| 0.8012 | 0.0 | 19 |
| 0.7568 | 0.0 | 20 |
| 0.7120 | 0.0 | 21 |
| 0.6746 | 0.0 | 22 |
| 0.6478 | 0.0 | 23 |
### Framework versions
- Transformers 4.35.2
- TensorFlow 2.15.0
- Datasets 2.16.1
- Tokenizers 0.15.0
|