dcampanini
commited on
Commit
•
c1f148b
1
Parent(s):
fc8a211
add a full version of the readme.d
Browse files
README.md
CHANGED
@@ -1,3 +1,52 @@
|
|
1 |
---
|
2 |
license: unknown
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: unknown
|
3 |
---
|
4 |
+
# LLaVA1.5-BiomedCLIP-Vicuna-7b for multimodal radiology report generation
|
5 |
+
|
6 |
+
This is a model based on LLaVA1.5-Vicuna-7b, finetuned to generate medical reports, based on a chest X-ray and a prompt, in our case, the instruction was "write the finding section of a chest x-ray radiology report".
|
7 |
+
|
8 |
+
The vision-encoder of the model is a [BiomedCLIP](https://huggingface.co/microsoft/BiomedCLIP-PubMedBERT_256-vit_base_patch16_224), the conector is a 2 layer MLP and the LLM is a Vicuna-7b-1.5v
|
9 |
+
|
10 |
+
The dataset used for finetuning was the MIMIC-CXR shared for the challenge in Radiology Report Generation for the Association for Computational Linguistics 2024 at BioNLP Workshop.
|
11 |
+
|
12 |
+
We used the 148,374 findings of MIMIC-CXR for finetuning during 3 epochs.
|
13 |
+
|
14 |
+
The model metrics on the 1,063 samples of the hidden test set of the ACL challenge are the following:
|
15 |
+
|
16 |
+
| Method | BLEU-4 | Rouge-L | Bertscore | F1-CheXbert | F1-RadGraph | Avg |
|
17 |
+
|-------------------------------|--------|---------|-----------|-------------|-------------|-------|
|
18 |
+
| llava1.5-biomedclip-Vicuna-7b | 3.48 | 16.31 | 35.49 | 29.37 | 15.51 | 20.03 |
|
19 |
+
|
20 |
+
|
21 |
+
The metrics were calculated direcly by the challenge organizer, however you can reproduce them with the following
|
22 |
+
example code:
|
23 |
+
|
24 |
+
```python
|
25 |
+
import json
|
26 |
+
import logging
|
27 |
+
from vilmedic.blocks.scorers.scores import compute_scores
|
28 |
+
|
29 |
+
refs = [
|
30 |
+
"The lungs are clear. The cardiomediastinal silhouette is within normal limits. No acute osseous abnormalities.",
|
31 |
+
"The lungs are clear.There is no pleural effusion or pneumothorax.The cardiomediastinal silhouette is normal."
|
32 |
+
]
|
33 |
+
hyps = [
|
34 |
+
"The lungs are clear. There is no pleural effusion or pneumothorax. The cardiomediastinal silhouette is normal.",
|
35 |
+
"The lungs are clear. The cardiomediastinal silhouette is within normal limits. No acute osseous abnormalities."
|
36 |
+
]
|
37 |
+
print("Computing metrics, this can take a while...")
|
38 |
+
print(json.dumps(compute_scores(["ROUGEL", "bertscore", "radgraph", "BLEU", "chexbert"],
|
39 |
+
refs=refs,
|
40 |
+
hyps=hyps,
|
41 |
+
split=None,
|
42 |
+
seed=None,
|
43 |
+
config=None,
|
44 |
+
epoch=None,
|
45 |
+
logger=logging.getLogger(__name__),
|
46 |
+
dump=False),
|
47 |
+
indent=4)
|
48 |
+
)
|
49 |
+
```
|
50 |
+
|
51 |
+
More details of the challenge can be found on the [challenge web page](https://stanford-aimi.github.io/RRG24/)
|
52 |
+
or in [workshop site](https://aclweb.org/aclwiki/BioNLP_Workshop)
|