model update
Browse files- README.md +2 -3
- config.json +1 -1
- pytorch_model.bin +2 -2
- tokenizer_config.json +1 -1
README.md
CHANGED
@@ -88,10 +88,10 @@ output = pipe('question: What is a person called is practicing heresy?, context:
|
|
88 |
|
89 |
```
|
90 |
|
91 |
-
## Evaluation
|
92 |
|
93 |
|
94 |
-
|
95 |
|
96 |
| | Score | Type | Dataset |
|
97 |
|:-----------------|--------:|:--------|:-------------------------------------------------------------------|
|
@@ -106,7 +106,6 @@ output = pipe('question: What is a person called is practicing heresy?, context:
|
|
106 |
| MoverScore | 80.07 | default | [lmqg/qg_tweetqa](https://huggingface.co/datasets/lmqg/qg_tweetqa) |
|
107 |
| ROUGE_L | 62.24 | default | [lmqg/qg_tweetqa](https://huggingface.co/datasets/lmqg/qg_tweetqa) |
|
108 |
|
109 |
-
- [raw metric file](https://huggingface.co/lmqg/t5-base-tweetqa-question-answering/raw/main/eval/metric.first.answer.paragraph_question.answer.lmqg_qg_tweetqa.default.json)
|
110 |
|
111 |
|
112 |
## Training hyperparameters
|
|
|
88 |
|
89 |
```
|
90 |
|
91 |
+
## Evaluation
|
92 |
|
93 |
|
94 |
+
- ***Metric (Question Answering)***: [raw metric file](https://huggingface.co/lmqg/t5-base-tweetqa-question-answering/raw/main/eval/metric.first.answer.paragraph_question.answer.lmqg_qg_tweetqa.default.json)
|
95 |
|
96 |
| | Score | Type | Dataset |
|
97 |
|:-----------------|--------:|:--------|:-------------------------------------------------------------------|
|
|
|
106 |
| MoverScore | 80.07 | default | [lmqg/qg_tweetqa](https://huggingface.co/datasets/lmqg/qg_tweetqa) |
|
107 |
| ROUGE_L | 62.24 | default | [lmqg/qg_tweetqa](https://huggingface.co/datasets/lmqg/qg_tweetqa) |
|
108 |
|
|
|
109 |
|
110 |
|
111 |
## Training hyperparameters
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "lmqg_output/t5-base-tweetqa-question-answering/
|
3 |
"add_prefix": false,
|
4 |
"architectures": [
|
5 |
"T5ForConditionalGeneration"
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "lmqg_output/t5-base-tweetqa-question-answering/model_eszyci/epoch_9",
|
3 |
"add_prefix": false,
|
4 |
"architectures": [
|
5 |
"T5ForConditionalGeneration"
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3d9af18f30d344085a0c0eb5fce63dbe9354c80444d241743054f9bc72089d2f
|
3 |
+
size 891617855
|
tokenizer_config.json
CHANGED
@@ -104,7 +104,7 @@
|
|
104 |
"eos_token": "</s>",
|
105 |
"extra_ids": 100,
|
106 |
"model_max_length": 512,
|
107 |
-
"name_or_path": "lmqg_output/t5-base-tweetqa-question-answering/
|
108 |
"pad_token": "<pad>",
|
109 |
"special_tokens_map_file": null,
|
110 |
"tokenizer_class": "T5Tokenizer",
|
|
|
104 |
"eos_token": "</s>",
|
105 |
"extra_ids": 100,
|
106 |
"model_max_length": 512,
|
107 |
+
"name_or_path": "lmqg_output/t5-base-tweetqa-question-answering/model_eszyci/epoch_9",
|
108 |
"pad_token": "<pad>",
|
109 |
"special_tokens_map_file": null,
|
110 |
"tokenizer_class": "T5Tokenizer",
|