End of training
Browse files- .gitattributes +1 -0
- README.md +51 -0
- config.json +37 -0
- generation_config.json +10 -0
- model.safetensors +3 -0
- runs/Jul22_20-27-51_1acbbaf73417/events.out.tfevents.1721680167.1acbbaf73417.278.0 +3 -0
- runs/Jul22_20-51-51_1acbbaf73417/events.out.tfevents.1721681688.1acbbaf73417.278.1 +3 -0
- runs/Jul22_20-51-51_1acbbaf73417/events.out.tfevents.1721683025.1acbbaf73417.278.2 +3 -0
- runs/Jul22_20-51-51_1acbbaf73417/events.out.tfevents.1721683083.1acbbaf73417.278.3 +3 -0
- runs/Jul22_22-22-32_1acbbaf73417/events.out.tfevents.1721687106.1acbbaf73417.31526.0 +3 -0
- runs/Jul22_22-22-32_1acbbaf73417/events.out.tfevents.1721695168.1acbbaf73417.31526.1 +3 -0
- special_tokens_map.json +23 -0
- spiece.model +3 -0
- tokenizer.json +3 -0
- tokenizer_config.json +41 -0
- training_args.bin +3 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model: ahmeddbahaa/mT5_multilingual_XLSum-finetuned-ar
|
3 |
+
tags:
|
4 |
+
- generated_from_trainer
|
5 |
+
model-index:
|
6 |
+
- name: results_mt5_xl-sum
|
7 |
+
results: []
|
8 |
+
---
|
9 |
+
|
10 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
11 |
+
should probably proofread and complete it, then remove this comment. -->
|
12 |
+
|
13 |
+
# results_mt5_xl-sum
|
14 |
+
|
15 |
+
This model is a fine-tuned version of [ahmeddbahaa/mT5_multilingual_XLSum-finetuned-ar](https://huggingface.co/ahmeddbahaa/mT5_multilingual_XLSum-finetuned-ar) on an unknown dataset.
|
16 |
+
|
17 |
+
## Model description
|
18 |
+
|
19 |
+
More information needed
|
20 |
+
|
21 |
+
## Intended uses & limitations
|
22 |
+
|
23 |
+
More information needed
|
24 |
+
|
25 |
+
## Training and evaluation data
|
26 |
+
|
27 |
+
More information needed
|
28 |
+
|
29 |
+
## Training procedure
|
30 |
+
|
31 |
+
### Training hyperparameters
|
32 |
+
|
33 |
+
The following hyperparameters were used during training:
|
34 |
+
- learning_rate: 0.0005
|
35 |
+
- train_batch_size: 2
|
36 |
+
- eval_batch_size: 2
|
37 |
+
- seed: 42
|
38 |
+
- gradient_accumulation_steps: 8
|
39 |
+
- total_train_batch_size: 16
|
40 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
41 |
+
- lr_scheduler_type: linear
|
42 |
+
- lr_scheduler_warmup_steps: 250
|
43 |
+
- num_epochs: 10
|
44 |
+
- mixed_precision_training: Native AMP
|
45 |
+
|
46 |
+
### Framework versions
|
47 |
+
|
48 |
+
- Transformers 4.40.0
|
49 |
+
- Pytorch 2.3.1+cu121
|
50 |
+
- Datasets 2.20.0
|
51 |
+
- Tokenizers 0.19.1
|
config.json
ADDED
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "ahmeddbahaa/mT5_multilingual_XLSum-finetuned-ar",
|
3 |
+
"architectures": [
|
4 |
+
"MT5ForConditionalGeneration"
|
5 |
+
],
|
6 |
+
"classifier_dropout": 0.0,
|
7 |
+
"d_ff": 2048,
|
8 |
+
"d_kv": 64,
|
9 |
+
"d_model": 768,
|
10 |
+
"decoder_start_token_id": 0,
|
11 |
+
"dense_act_fn": "gelu_new",
|
12 |
+
"dropout_rate": 0.1,
|
13 |
+
"eos_token_id": 1,
|
14 |
+
"feed_forward_proj": "gated-gelu",
|
15 |
+
"initializer_factor": 1.0,
|
16 |
+
"is_encoder_decoder": true,
|
17 |
+
"is_gated_act": true,
|
18 |
+
"layer_norm_epsilon": 1e-06,
|
19 |
+
"length_penalty": 0.6,
|
20 |
+
"max_length": 84,
|
21 |
+
"model_type": "mt5",
|
22 |
+
"no_repeat_ngram_size": 2,
|
23 |
+
"num_beams": 4,
|
24 |
+
"num_decoder_layers": 12,
|
25 |
+
"num_heads": 12,
|
26 |
+
"num_layers": 12,
|
27 |
+
"output_past": true,
|
28 |
+
"pad_token_id": 0,
|
29 |
+
"relative_attention_max_distance": 128,
|
30 |
+
"relative_attention_num_buckets": 32,
|
31 |
+
"tie_word_embeddings": false,
|
32 |
+
"tokenizer_class": "T5Tokenizer",
|
33 |
+
"torch_dtype": "float32",
|
34 |
+
"transformers_version": "4.40.0",
|
35 |
+
"use_cache": true,
|
36 |
+
"vocab_size": 250112
|
37 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"decoder_start_token_id": 0,
|
3 |
+
"eos_token_id": 1,
|
4 |
+
"length_penalty": 0.6,
|
5 |
+
"max_length": 84,
|
6 |
+
"no_repeat_ngram_size": 2,
|
7 |
+
"num_beams": 4,
|
8 |
+
"pad_token_id": 0,
|
9 |
+
"transformers_version": "4.40.0"
|
10 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2f0df86a0b364ee3ebf8cc7ee044de831c60efb2b16c4adb3b6c63290f97727a
|
3 |
+
size 2329601904
|
runs/Jul22_20-27-51_1acbbaf73417/events.out.tfevents.1721680167.1acbbaf73417.278.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8798012bc47e567c63a6114e1dfb05d489fca0ed32b0376fc40fcd5d7fa8d093
|
3 |
+
size 14316
|
runs/Jul22_20-51-51_1acbbaf73417/events.out.tfevents.1721681688.1acbbaf73417.278.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1f717fcb459bf58f978ab78d642d34f6578d7de59991fde6de2994b0b39b0ba9
|
3 |
+
size 7202
|
runs/Jul22_20-51-51_1acbbaf73417/events.out.tfevents.1721683025.1acbbaf73417.278.2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:189342f0db5539ac87eb6aeea0bcdcd50bd6158f996ad4d2682f5f66e9097fcd
|
3 |
+
size 7205
|
runs/Jul22_20-51-51_1acbbaf73417/events.out.tfevents.1721683083.1acbbaf73417.278.3
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0a3371aa3ff78d6a4ba30bb8efe098c1128b08b6d8b5a3d2ed73efd965b6234b
|
3 |
+
size 7202
|
runs/Jul22_22-22-32_1acbbaf73417/events.out.tfevents.1721687106.1acbbaf73417.31526.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:af381a48e0ca257ed4d1e75f81470971848c8d6e71f0eadf071fe0a5d4b7e059
|
3 |
+
size 6764
|
runs/Jul22_22-22-32_1acbbaf73417/events.out.tfevents.1721695168.1acbbaf73417.31526.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3fa2882c0147273f12f2d0799e9ad97536f866e4a4f3dc7f23826dec7bbe5393
|
3 |
+
size 6764
|
special_tokens_map.json
ADDED
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eos_token": {
|
3 |
+
"content": "</s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"pad_token": {
|
10 |
+
"content": "<pad>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"unk_token": {
|
17 |
+
"content": "<unk>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
}
|
23 |
+
}
|
spiece.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ef78f86560d809067d12bac6c09f19a462cb3af3f54d2b8acbba26e1433125d6
|
3 |
+
size 4309802
|
tokenizer.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:59a46e92c70ae431bcb2de7cf875d9cf2a4eecd1a41c9ed516208800929118a1
|
3 |
+
size 16315456
|
tokenizer_config.json
ADDED
@@ -0,0 +1,41 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "<pad>",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "</s>",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"2": {
|
20 |
+
"content": "<unk>",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
}
|
27 |
+
},
|
28 |
+
"additional_special_tokens": [],
|
29 |
+
"clean_up_tokenization_spaces": true,
|
30 |
+
"eos_token": "</s>",
|
31 |
+
"extra_ids": 0,
|
32 |
+
"max_length": 128,
|
33 |
+
"model_max_length": 1000000000000000019884624838656,
|
34 |
+
"pad_token": "<pad>",
|
35 |
+
"sp_model_kwargs": {},
|
36 |
+
"stride": 0,
|
37 |
+
"tokenizer_class": "T5Tokenizer",
|
38 |
+
"truncation_side": "right",
|
39 |
+
"truncation_strategy": "longest_first",
|
40 |
+
"unk_token": "<unk>"
|
41 |
+
}
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4fda699c241d0c8c9bd3a8265561b86c476853abb0ce15bb55d1b1b3eeccfb7b
|
3 |
+
size 6712
|