w11wo commited on
Commit
63b6ad1
1 Parent(s): 8066982

End of training

Browse files
README.md ADDED
@@ -0,0 +1,90 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: LazarusNLP/IndoNanoT5-base
4
+ tags:
5
+ - generated_from_trainer
6
+ datasets:
7
+ - indonlg
8
+ metrics:
9
+ - bleu
10
+ - sacrebleu
11
+ model-index:
12
+ - name: IndoNanoT5-base-XPersona
13
+ results:
14
+ - task:
15
+ name: Sequence-to-sequence Language Modeling
16
+ type: text2text-generation
17
+ dataset:
18
+ name: indonlg
19
+ type: indonlg
20
+ config: xpersona
21
+ split: validation
22
+ args: xpersona
23
+ metrics:
24
+ - name: Bleu
25
+ type: bleu
26
+ value: 4.0669
27
+ - name: Sacrebleu
28
+ type: sacrebleu
29
+ value: 4.0669
30
+ ---
31
+
32
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
33
+ should probably proofread and complete it, then remove this comment. -->
34
+
35
+ # IndoNanoT5-base-XPersona
36
+
37
+ This model is a fine-tuned version of [LazarusNLP/IndoNanoT5-base](https://huggingface.co/LazarusNLP/IndoNanoT5-base) on the indonlg dataset.
38
+ It achieves the following results on the evaluation set:
39
+ - Loss: 1.8372
40
+ - Bleu: 4.0669
41
+ - Sacrebleu: 4.0669
42
+
43
+ ## Model description
44
+
45
+ More information needed
46
+
47
+ ## Intended uses & limitations
48
+
49
+ More information needed
50
+
51
+ ## Training and evaluation data
52
+
53
+ More information needed
54
+
55
+ ## Training procedure
56
+
57
+ ### Training hyperparameters
58
+
59
+ The following hyperparameters were used during training:
60
+ - learning_rate: 1e-05
61
+ - train_batch_size: 8
62
+ - eval_batch_size: 16
63
+ - seed: 42
64
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
65
+ - lr_scheduler_type: linear
66
+ - num_epochs: 50
67
+
68
+ ### Training results
69
+
70
+ | Training Loss | Epoch | Step | Validation Loss | Bleu | Sacrebleu |
71
+ |:-------------:|:-----:|:------:|:---------------:|:------:|:---------:|
72
+ | 1.9872 | 1.0 | 15516 | 1.8482 | 3.7015 | 3.7015 |
73
+ | 1.888 | 2.0 | 31032 | 1.8434 | 4.0409 | 4.0409 |
74
+ | 1.8207 | 3.0 | 46548 | 1.8347 | 4.1239 | 4.1239 |
75
+ | 1.7716 | 4.0 | 62064 | 1.8340 | 4.3231 | 4.3231 |
76
+ | 1.6948 | 5.0 | 77580 | 1.8443 | 4.4283 | 4.4283 |
77
+ | 1.6442 | 6.0 | 93096 | 1.8563 | 4.5338 | 4.5338 |
78
+ | 1.5856 | 7.0 | 108612 | 1.8782 | 4.3033 | 4.3033 |
79
+ | 1.5451 | 8.0 | 124128 | 1.8930 | 4.3286 | 4.3286 |
80
+ | 1.5056 | 9.0 | 139644 | 1.9207 | 4.2773 | 4.2773 |
81
+ | 1.446 | 10.0 | 155160 | 1.9406 | 4.0629 | 4.0629 |
82
+ | 1.406 | 11.0 | 170676 | 1.9636 | 4.1382 | 4.1382 |
83
+
84
+
85
+ ### Framework versions
86
+
87
+ - Transformers 4.37.2
88
+ - Pytorch 2.2.0+cu118
89
+ - Datasets 2.16.1
90
+ - Tokenizers 0.15.1
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "decoder_start_token_id": 0,
3
+ "eos_token_id": 1,
4
+ "pad_token_id": 0,
5
+ "transformers_version": "4.37.2"
6
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d47a6fece82b84b29e9ab4fc1ad0a48c88be6f4b26066bf2b3a32d25bd0ddcfc
3
  size 990345064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a401ce049659acb59b1854cfadec4db33589f6fb41aeac38ff80d4ca5c989994
3
  size 990345064
runs/Feb09_07-23-37_bookbot-h100/events.out.tfevents.1707478684.bookbot-h100 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8157db4c261a8192762a8530d59288c66a03f9a725cca367046da68343e643da
3
+ size 40