phoner45 commited on
Commit
65802af
1 Parent(s): fadf721

Training complete

Browse files
README.md CHANGED
@@ -1,8 +1,11 @@
1
  ---
2
  library_name: transformers
3
- base_model: airesearch/wangchanberta-base-att-spm-uncased
 
4
  tags:
5
  - generated_from_trainer
 
 
6
  model-index:
7
  - name: finetune_iapp_thaiqa_v2
8
  results: []
@@ -13,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # finetune_iapp_thaiqa_v2
15
 
16
- This model is a fine-tuned version of [airesearch/wangchanberta-base-att-spm-uncased](https://huggingface.co/airesearch/wangchanberta-base-att-spm-uncased) on the None dataset.
17
 
18
  ## Model description
19
 
 
1
  ---
2
  library_name: transformers
3
+ license: mit
4
+ base_model: phoner45/phoner45-finetune-iapp-thaiqa
5
  tags:
6
  - generated_from_trainer
7
+ datasets:
8
+ - iapp_wiki_qa_squad
9
  model-index:
10
  - name: finetune_iapp_thaiqa_v2
11
  results: []
 
16
 
17
  # finetune_iapp_thaiqa_v2
18
 
19
+ This model is a fine-tuned version of [phoner45/phoner45-finetune-iapp-thaiqa](https://huggingface.co/phoner45/phoner45-finetune-iapp-thaiqa) on the iapp_wiki_qa_squad dataset.
20
 
21
  ## Model description
22
 
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "airesearch/wangchanberta-base-att-spm-uncased",
3
  "architectures": [
4
  "CamembertForQuestionAnswering"
5
  ],
 
1
  {
2
+ "_name_or_path": "phoner45/phoner45-finetune-iapp-thaiqa",
3
  "architectures": [
4
  "CamembertForQuestionAnswering"
5
  ],
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5214a9e74c44ac291c81c11f1a74d62548857e698c6f63b836d06f89b18d2056
3
  size 418645336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01b4cfab97b262eb19a397708eea8203550cb068ffcd013adecf050bba57ced6
3
  size 418645336
sentencepiece.bpe.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:49c4ba4e495ddf31eb2fdba7fc6aef3c233091d25d35bc9d24694ccf48ae114c
3
- size 904693
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2059a7452cb4f0254070b2d4f810daf1027dda9329c3f8bbadfa2bcc55a8282e
3
+ size 872531
special_tokens_map.json CHANGED
@@ -2,19 +2,55 @@
2
  "additional_special_tokens": [
3
  "<s>NOTUSED",
4
  "</s>NOTUSED",
5
- "<_>"
6
  ],
7
- "bos_token": "<s>",
8
- "cls_token": "<s>",
9
- "eos_token": "</s>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  "mask_token": {
11
  "content": "<mask>",
12
  "lstrip": true,
 
 
 
 
 
 
 
13
  "normalized": false,
14
  "rstrip": false,
15
  "single_word": false
16
  },
17
- "pad_token": "<pad>",
18
- "sep_token": "</s>",
19
- "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
20
  }
 
2
  "additional_special_tokens": [
3
  "<s>NOTUSED",
4
  "</s>NOTUSED",
5
+ ""
6
  ],
7
+ "bos_token": {
8
+ "content": "<s>",
9
+ "lstrip": false,
10
+ "normalized": false,
11
+ "rstrip": false,
12
+ "single_word": false
13
+ },
14
+ "cls_token": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "eos_token": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false
27
+ },
28
  "mask_token": {
29
  "content": "<mask>",
30
  "lstrip": true,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false
34
+ },
35
+ "pad_token": {
36
+ "content": "<pad>",
37
+ "lstrip": false,
38
  "normalized": false,
39
  "rstrip": false,
40
  "single_word": false
41
  },
42
+ "sep_token": {
43
+ "content": "</s>",
44
+ "lstrip": false,
45
+ "normalized": false,
46
+ "rstrip": false,
47
+ "single_word": false
48
+ },
49
+ "unk_token": {
50
+ "content": "<unk>",
51
+ "lstrip": false,
52
+ "normalized": false,
53
+ "rstrip": false,
54
+ "single_word": false
55
+ }
56
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -49,17 +49,17 @@
49
  "special": true
50
  },
51
  "8": {
52
- "content": "<_>",
53
  "lstrip": false,
54
  "normalized": false,
55
  "rstrip": false,
56
  "single_word": false,
57
- "special": false
58
  },
59
- "25004": {
60
  "content": "<mask>",
61
  "lstrip": true,
62
- "normalized": false,
63
  "rstrip": false,
64
  "single_word": false,
65
  "special": true
@@ -68,17 +68,24 @@
68
  "additional_special_tokens": [
69
  "<s>NOTUSED",
70
  "</s>NOTUSED",
71
- "<_>"
72
  ],
73
  "bos_token": "<s>",
74
  "clean_up_tokenization_spaces": true,
75
  "cls_token": "<s>",
76
  "eos_token": "</s>",
77
  "mask_token": "<mask>",
78
- "model_max_length": 416,
 
 
79
  "pad_token": "<pad>",
 
 
80
  "sep_token": "</s>",
81
  "sp_model_kwargs": {},
 
82
  "tokenizer_class": "CamembertTokenizer",
 
 
83
  "unk_token": "<unk>"
84
  }
 
49
  "special": true
50
  },
51
  "8": {
52
+ "content": "",
53
  "lstrip": false,
54
  "normalized": false,
55
  "rstrip": false,
56
  "single_word": false,
57
+ "special": true
58
  },
59
+ "24004": {
60
  "content": "<mask>",
61
  "lstrip": true,
62
+ "normalized": true,
63
  "rstrip": false,
64
  "single_word": false,
65
  "special": true
 
68
  "additional_special_tokens": [
69
  "<s>NOTUSED",
70
  "</s>NOTUSED",
71
+ ""
72
  ],
73
  "bos_token": "<s>",
74
  "clean_up_tokenization_spaces": true,
75
  "cls_token": "<s>",
76
  "eos_token": "</s>",
77
  "mask_token": "<mask>",
78
+ "max_length": 416,
79
+ "model_max_length": 400,
80
+ "pad_to_multiple_of": null,
81
  "pad_token": "<pad>",
82
+ "pad_token_type_id": 0,
83
+ "padding_side": "right",
84
  "sep_token": "</s>",
85
  "sp_model_kwargs": {},
86
+ "stride": 128,
87
  "tokenizer_class": "CamembertTokenizer",
88
+ "truncation_side": "right",
89
+ "truncation_strategy": "only_second",
90
  "unk_token": "<unk>"
91
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c04b6e606152ff67ac5881cfe265a8297b112e0183d2d74edb3dfd95aa53ddc
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11293ab700d07db430f5c5f77d3efb0d3eeb651109642d76b2b2bad885c65ba6
3
  size 5176