Chenxi Whitehouse
commited on
Commit
•
857a089
1
Parent(s):
eca41fc
update readme
Browse files
README.md
CHANGED
@@ -41,6 +41,7 @@ The training and dev dataset can be found under [data](https://huggingface.co/ch
|
|
41 |
}
|
42 |
]
|
43 |
},
|
|
|
44 |
}
|
45 |
```
|
46 |
|
@@ -98,13 +99,13 @@ python -m src.reranking.question_generation_top_sentences
|
|
98 |
```
|
99 |
|
100 |
### 4. Rerank the QA pairs
|
101 |
-
Using
|
102 |
```bash
|
103 |
```
|
104 |
|
105 |
|
106 |
### 5. Veracity prediction
|
107 |
-
Finally, given a claim and its 3 QA pairs as evidence, we use
|
108 |
```bash
|
109 |
```
|
110 |
The results will be presented as follows:
|
|
|
41 |
}
|
42 |
]
|
43 |
},
|
44 |
+
]
|
45 |
}
|
46 |
```
|
47 |
|
|
|
99 |
```
|
100 |
|
101 |
### 4. Rerank the QA pairs
|
102 |
+
Using a pre-trained BERT model [bert_dual_encoder.ckpt](https://huggingface.co/chenxwh/AVeriTeC/blob/main/pretrained_models/bert_dual_encoder.ckpt) we rerank the QA paris and keep top 3 QA paris as evidence. We provide the output file for this step on the dev set [here]().
|
103 |
```bash
|
104 |
```
|
105 |
|
106 |
|
107 |
### 5. Veracity prediction
|
108 |
+
Finally, given a claim and its 3 QA pairs as evidence, we use another pre-trained BERT model [bert_veracity.ckpt](https://huggingface.co/chenxwh/AVeriTeC/blob/main/pretrained_models/bert_veracity.ckpt) to predict the veracity label. The pre-trained model is provided . We provide the prediction file for this step on the dev set [here]().
|
109 |
```bash
|
110 |
```
|
111 |
The results will be presented as follows:
|