metadata
license: apache-2.0
base_model: hustvl/yolos-tiny
tags:
- generated_from_trainer
model-index:
- name: Yolo_test
results: []
Yolo_test
This model is a fine-tuned version of hustvl/yolos-tiny on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 0.9194
- Map: 0.1092
- Map 50: 0.2705
- Map 75: 0.0385
- Map Small: 0.2009
- Map Medium: 0.0867
- Map Large: -1.0
- Mar 1: 0.1037
- Mar 10: 0.4296
- Mar 100: 0.5222
- Mar Small: 0.61
- Mar Medium: 0.4706
- Mar Large: -1.0
- Map Background : -1.0
- Mar 100 Background : -1.0
- Map Score: 0.1092
- Mar 100 Score: 0.5222
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- num_epochs: 30
Training results
Training Loss | Epoch | Step | Validation Loss | Map | Map 50 | Map 75 | Map Small | Map Medium | Map Large | Mar 1 | Mar 10 | Mar 100 | Mar Small | Mar Medium | Mar Large | Map Background | Mar 100 Background | Map Score | Mar 100 Score |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
No log | 1.0 | 93 | 1.2349 | 0.0104 | 0.0497 | 0.0004 | 0.0266 | 0.0053 | -1.0 | 0.0333 | 0.1741 | 0.2556 | 0.44 | 0.1471 | -1.0 | -1.0 | -1.0 | 0.0104 | 0.2556 |
No log | 2.0 | 186 | 1.3555 | 0.0071 | 0.038 | 0.001 | 0.0205 | 0.0032 | -1.0 | 0.0037 | 0.1444 | 0.1704 | 0.24 | 0.1294 | -1.0 | -1.0 | -1.0 | 0.0071 | 0.1704 |
No log | 3.0 | 279 | 1.3345 | 0.0224 | 0.0669 | 0.0048 | 0.0557 | 0.0124 | -1.0 | 0.063 | 0.1926 | 0.2148 | 0.29 | 0.1706 | -1.0 | -1.0 | -1.0 | 0.0224 | 0.2148 |
No log | 4.0 | 372 | 1.1534 | 0.0709 | 0.2465 | 0.0068 | 0.1562 | 0.0233 | -1.0 | 0.1 | 0.3111 | 0.3259 | 0.46 | 0.2471 | -1.0 | -1.0 | -1.0 | 0.0709 | 0.3259 |
No log | 5.0 | 465 | 1.0287 | 0.0536 | 0.1696 | 0.0025 | 0.1018 | 0.0286 | -1.0 | 0.063 | 0.3259 | 0.3444 | 0.37 | 0.3294 | -1.0 | -1.0 | -1.0 | 0.0536 | 0.3444 |
1.2502 | 6.0 | 558 | 1.0684 | 0.0381 | 0.1233 | 0.0015 | 0.0979 | 0.0184 | -1.0 | 0.0593 | 0.2889 | 0.3556 | 0.44 | 0.3059 | -1.0 | -1.0 | -1.0 | 0.0381 | 0.3556 |
1.2502 | 7.0 | 651 | 1.0613 | 0.0252 | 0.0884 | 0.0024 | 0.0409 | 0.0208 | -1.0 | 0.0407 | 0.2593 | 0.3407 | 0.35 | 0.3353 | -1.0 | -1.0 | -1.0 | 0.0252 | 0.3407 |
1.2502 | 8.0 | 744 | 0.9724 | 0.0602 | 0.2019 | 0.008 | 0.1258 | 0.034 | -1.0 | 0.1407 | 0.3704 | 0.4296 | 0.5 | 0.3882 | -1.0 | -1.0 | -1.0 | 0.0602 | 0.4296 |
1.2502 | 9.0 | 837 | 1.0170 | 0.055 | 0.1548 | 0.0077 | 0.0627 | 0.0556 | -1.0 | 0.0667 | 0.3889 | 0.3889 | 0.41 | 0.3765 | -1.0 | -1.0 | -1.0 | 0.055 | 0.3889 |
1.2502 | 10.0 | 930 | 1.1431 | 0.0324 | 0.1301 | 0.0025 | 0.0532 | 0.0281 | -1.0 | 0.063 | 0.2741 | 0.3037 | 0.32 | 0.2941 | -1.0 | -1.0 | -1.0 | 0.0324 | 0.3037 |
0.8688 | 11.0 | 1023 | 0.9432 | 0.0511 | 0.1614 | 0.0152 | 0.1059 | 0.0394 | -1.0 | 0.0333 | 0.3407 | 0.4481 | 0.49 | 0.4235 | -1.0 | -1.0 | -1.0 | 0.0511 | 0.4481 |
0.8688 | 12.0 | 1116 | 0.9395 | 0.0658 | 0.1824 | 0.0115 | 0.1559 | 0.0409 | -1.0 | 0.0815 | 0.3741 | 0.4333 | 0.49 | 0.4 | -1.0 | -1.0 | -1.0 | 0.0658 | 0.4333 |
0.8688 | 13.0 | 1209 | 1.0367 | 0.0554 | 0.1926 | 0.0138 | 0.123 | 0.0377 | -1.0 | 0.1074 | 0.3 | 0.3667 | 0.47 | 0.3059 | -1.0 | -1.0 | -1.0 | 0.0554 | 0.3667 |
0.8688 | 14.0 | 1302 | 1.0167 | 0.0459 | 0.1409 | 0.0077 | 0.0799 | 0.0452 | -1.0 | 0.0444 | 0.3 | 0.4 | 0.46 | 0.3647 | -1.0 | -1.0 | -1.0 | 0.0459 | 0.4 |
0.8688 | 15.0 | 1395 | 1.0687 | 0.0403 | 0.126 | 0.005 | 0.0494 | 0.0472 | -1.0 | 0.0778 | 0.2741 | 0.3444 | 0.48 | 0.2647 | -1.0 | -1.0 | -1.0 | 0.0403 | 0.3444 |
0.8688 | 16.0 | 1488 | 0.9219 | 0.1075 | 0.2652 | 0.0661 | 0.233 | 0.0563 | -1.0 | 0.1185 | 0.3741 | 0.4889 | 0.58 | 0.4353 | -1.0 | -1.0 | -1.0 | 0.1075 | 0.4889 |
0.7025 | 17.0 | 1581 | 0.9246 | 0.0706 | 0.2286 | 0.0293 | 0.2273 | 0.0339 | -1.0 | 0.1111 | 0.3778 | 0.4444 | 0.58 | 0.3647 | -1.0 | -1.0 | -1.0 | 0.0706 | 0.4444 |
0.7025 | 18.0 | 1674 | 0.9737 | 0.0638 | 0.1656 | 0.0298 | 0.1258 | 0.0524 | -1.0 | 0.0778 | 0.4111 | 0.463 | 0.55 | 0.4118 | -1.0 | -1.0 | -1.0 | 0.0638 | 0.463 |
0.7025 | 19.0 | 1767 | 0.9821 | 0.0687 | 0.2266 | 0.0189 | 0.1069 | 0.0594 | -1.0 | 0.0444 | 0.3704 | 0.4259 | 0.46 | 0.4059 | -1.0 | -1.0 | -1.0 | 0.0687 | 0.4259 |
0.7025 | 20.0 | 1860 | 0.9198 | 0.0966 | 0.3077 | 0.0152 | 0.212 | 0.058 | -1.0 | 0.1519 | 0.3704 | 0.4778 | 0.59 | 0.4118 | -1.0 | -1.0 | -1.0 | 0.0966 | 0.4778 |
0.7025 | 21.0 | 1953 | 0.9468 | 0.087 | 0.2415 | 0.0305 | 0.1982 | 0.0708 | -1.0 | 0.163 | 0.3778 | 0.4667 | 0.51 | 0.4412 | -1.0 | -1.0 | -1.0 | 0.087 | 0.4667 |
0.5994 | 22.0 | 2046 | 0.8969 | 0.1079 | 0.2414 | 0.0639 | 0.2735 | 0.0765 | -1.0 | 0.1444 | 0.4296 | 0.5407 | 0.64 | 0.4824 | -1.0 | -1.0 | -1.0 | 0.1079 | 0.5407 |
0.5994 | 23.0 | 2139 | 0.9532 | 0.0974 | 0.233 | 0.0462 | 0.249 | 0.0641 | -1.0 | 0.163 | 0.3963 | 0.5111 | 0.62 | 0.4471 | -1.0 | -1.0 | -1.0 | 0.0974 | 0.5111 |
0.5994 | 24.0 | 2232 | 0.9191 | 0.104 | 0.2925 | 0.0506 | 0.2068 | 0.0671 | -1.0 | 0.1148 | 0.4 | 0.4926 | 0.62 | 0.4176 | -1.0 | -1.0 | -1.0 | 0.104 | 0.4926 |
0.5994 | 25.0 | 2325 | 0.8977 | 0.1095 | 0.3031 | 0.0364 | 0.1874 | 0.0843 | -1.0 | 0.1148 | 0.437 | 0.5296 | 0.63 | 0.4706 | -1.0 | -1.0 | -1.0 | 0.1095 | 0.5296 |
0.5994 | 26.0 | 2418 | 0.9240 | 0.0999 | 0.2688 | 0.0423 | 0.1689 | 0.0872 | -1.0 | 0.0963 | 0.4222 | 0.5185 | 0.59 | 0.4765 | -1.0 | -1.0 | -1.0 | 0.0999 | 0.5185 |
0.526 | 27.0 | 2511 | 0.9139 | 0.1091 | 0.2691 | 0.038 | 0.2088 | 0.0872 | -1.0 | 0.1259 | 0.4222 | 0.5185 | 0.6 | 0.4706 | -1.0 | -1.0 | -1.0 | 0.1091 | 0.5185 |
0.526 | 28.0 | 2604 | 0.9143 | 0.1058 | 0.2494 | 0.0547 | 0.2091 | 0.0827 | -1.0 | 0.1037 | 0.4296 | 0.5296 | 0.62 | 0.4765 | -1.0 | -1.0 | -1.0 | 0.1058 | 0.5296 |
0.526 | 29.0 | 2697 | 0.9166 | 0.1103 | 0.2704 | 0.037 | 0.2046 | 0.0867 | -1.0 | 0.1074 | 0.4259 | 0.5259 | 0.61 | 0.4765 | -1.0 | -1.0 | -1.0 | 0.1103 | 0.5259 |
0.526 | 30.0 | 2790 | 0.9194 | 0.1092 | 0.2705 | 0.0385 | 0.2009 | 0.0867 | -1.0 | 0.1037 | 0.4296 | 0.5222 | 0.61 | 0.4706 | -1.0 | -1.0 | -1.0 | 0.1092 | 0.5222 |
Framework versions
- Transformers 4.42.4
- Pytorch 2.3.1+cu121
- Datasets 2.20.0
- Tokenizers 0.19.1