|
--- |
|
license: apache-2.0 |
|
metrics: |
|
- mse |
|
- mae |
|
- r_squared |
|
tags: |
|
- medical |
|
model-index: |
|
- name: Vit-COVID-19-severity |
|
results: |
|
- task: |
|
name: Image Classification |
|
type: image-classification |
|
metrics: |
|
- name: MSE |
|
type: mse |
|
value: 0.5135470628738403 |
|
- name: MAE |
|
type: mae |
|
value: 0.5676153898239136 |
|
- name: R^ 2 |
|
type: r_squared |
|
value: 0.5378701686859131 |
|
pipeline_tag: image-classification |
|
--- |
|
|
|
## Diagnosing Medical Images for COVID-19 Severity (Regression task) |
|
|
|
The COVID-19 pandemic has strained healthcare resources and prompted discussion about how machine learning can alleviate physician burdens and contribute to diagnosis. Chest x-rays (CXRs) are used for diagnosis of COVID-19, but few studies predict the severity of a patient’s condition from CXRs. |
|
In this study, we produce a large COVID severity dataset by merging three sources and investigate the efficacy of transfer learning vision transformers (ViTs) in severity regression task. |
|
The ViT had the best regression results, with an MSE of 0.5135. Code developed in this project is available at https://github.com/stwhitfield/ift6759_project. |
|
|
|
## Model description |
|
|
|
We were fine-tuned the Vision Transformer (ViT) from HuggingFaces on our training set for 300 epochs using an AdamW optimizer with a learning rate of 0.00001 and a batch size of 10. |
|
Early stopping was applied to the model to prevent over-fitting. |
|
For ViT, training stopped after 70 epochs. The only difference between the preprocessed data in the ViT and the other models was that it normalized to a tensor image with mean and standard deviation during data augmentation. |