YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
Transfer Learning Vision Transformer (ViT) - Google 224 ViT Base Patch
Description
This model is a Transfer Learning Vision Transformer (ViT) based on Google's 224 ViT Base Patch architecture. It has been fine-tuned on a dataset consisting of fungal images from Russia, with a specific focus on various fungi and lichen species.
Model Information
- Model Name: Transfer Learning ViT - Google 224 ViT Base Patch
- Model Architecture: Vision Transformer (ViT)
- Base Architecture: Google's 224 ViT Base Patch
- Pre-trained on General ImageNet dataset
- Fine-tuned on: Fungal image dataset from Russia
Performance
- Accuracy: 90.31%
- F1 Score: 86.33%
Training Details
- Training Loss:
- Initial: 1.043200
- Final: 0.116200
- Validation Loss:
- Initial: 0.822428
- Final: 0.335994
- Training Epochs: 10
- Training Runtime: 18575.04 seconds
- Training Samples per Second: 33.327
- Training Steps per Second: 1.042
- Total FLOPs: 4.801 x 10^19
Recommended Use Cases
- Species classification of various fungi and lichen in Russia.
- Fungal biodiversity studies.
- Image recognition tasks related to fungi and lichen species.
Limitations
- The model's performance is optimized for fungal species and may not generalize well to other domains.
- The model may not perform well on images of fungi and lichen species from regions other than Russia.
Model Author
Siddhant Dutta
- Downloads last month
- 3
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.