File size: 2,879 Bytes
9b6b578
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
bca0245
9b6b578
bca0245
9b6b578
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9bcdbb5
bd725a7
4ed6107
cc8f52c
2117a6d
182964c
f0e6393
dee7e75
959b130
ed993f1
4b1283c
11ea75c
9d023ba
bb2948f
0ceb3ed
e45051a
93fdbea
0dd7b6b
3beff54
aa9992f
9b5a2c7
c8bded5
7320b69
c5f29ce
13c4374
7097451
a24bb2a
a06542d
bca0245
9b6b578
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
---
license: mit
base_model: badokorach/afriqa_afroxlmr_squad_v2-luganda
tags:
- generated_from_keras_callback
model-index:
- name: badokorach/afriqa_afroxlmr_squad_v2-luganda_311223
  results: []
---

<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->

# badokorach/afriqa_afroxlmr_squad_v2-luganda_311223

This model is a fine-tuned version of [badokorach/afriqa_afroxlmr_squad_v2-luganda](https://huggingface.co/badokorach/afriqa_afroxlmr_squad_v2-luganda) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 0.5410
- Validation Loss: 0.0
- Epoch: 29

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 1e-05, 'decay_steps': 14760, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.02}
- training_precision: mixed_float16

### Training results

| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 3.4699     | 0.0             | 0     |
| 3.3401     | 0.0             | 1     |
| 3.1487     | 0.0             | 2     |
| 2.9589     | 0.0             | 3     |
| 2.7472     | 0.0             | 4     |
| 2.5135     | 0.0             | 5     |
| 2.3000     | 0.0             | 6     |
| 2.0904     | 0.0             | 7     |
| 1.9314     | 0.0             | 8     |
| 1.7617     | 0.0             | 9     |
| 1.6075     | 0.0             | 10    |
| 1.4573     | 0.0             | 11    |
| 1.3275     | 0.0             | 12    |
| 1.2261     | 0.0             | 13    |
| 1.1378     | 0.0             | 14    |
| 1.0358     | 0.0             | 15    |
| 0.9618     | 0.0             | 16    |
| 0.9082     | 0.0             | 17    |
| 0.8762     | 0.0             | 18    |
| 0.8012     | 0.0             | 19    |
| 0.7568     | 0.0             | 20    |
| 0.7120     | 0.0             | 21    |
| 0.6746     | 0.0             | 22    |
| 0.6478     | 0.0             | 23    |
| 0.6286     | 0.0             | 24    |
| 0.5978     | 0.0             | 25    |
| 0.5822     | 0.0             | 26    |
| 0.5444     | 0.0             | 27    |
| 0.5344     | 0.0             | 28    |
| 0.5410     | 0.0             | 29    |


### Framework versions

- Transformers 4.35.2
- TensorFlow 2.15.0
- Datasets 2.16.1
- Tokenizers 0.15.0