Yolo_test / README.md
SmallPS's picture
End of training
fead735 verified
|
raw
history blame
9.53 kB
metadata
license: apache-2.0
base_model: hustvl/yolos-tiny
tags:
  - generated_from_trainer
model-index:
  - name: Yolo_test
    results: []

Yolo_test

This model is a fine-tuned version of hustvl/yolos-tiny on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9194
  • Map: 0.1092
  • Map 50: 0.2705
  • Map 75: 0.0385
  • Map Small: 0.2009
  • Map Medium: 0.0867
  • Map Large: -1.0
  • Mar 1: 0.1037
  • Mar 10: 0.4296
  • Mar 100: 0.5222
  • Mar Small: 0.61
  • Mar Medium: 0.4706
  • Mar Large: -1.0
  • Map Background : -1.0
  • Mar 100 Background : -1.0
  • Map Score: 0.1092
  • Mar 100 Score: 0.5222

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: cosine
  • num_epochs: 30

Training results

Training Loss Epoch Step Validation Loss Map Map 50 Map 75 Map Small Map Medium Map Large Mar 1 Mar 10 Mar 100 Mar Small Mar Medium Mar Large Map Background Mar 100 Background Map Score Mar 100 Score
No log 1.0 93 1.2349 0.0104 0.0497 0.0004 0.0266 0.0053 -1.0 0.0333 0.1741 0.2556 0.44 0.1471 -1.0 -1.0 -1.0 0.0104 0.2556
No log 2.0 186 1.3555 0.0071 0.038 0.001 0.0205 0.0032 -1.0 0.0037 0.1444 0.1704 0.24 0.1294 -1.0 -1.0 -1.0 0.0071 0.1704
No log 3.0 279 1.3345 0.0224 0.0669 0.0048 0.0557 0.0124 -1.0 0.063 0.1926 0.2148 0.29 0.1706 -1.0 -1.0 -1.0 0.0224 0.2148
No log 4.0 372 1.1534 0.0709 0.2465 0.0068 0.1562 0.0233 -1.0 0.1 0.3111 0.3259 0.46 0.2471 -1.0 -1.0 -1.0 0.0709 0.3259
No log 5.0 465 1.0287 0.0536 0.1696 0.0025 0.1018 0.0286 -1.0 0.063 0.3259 0.3444 0.37 0.3294 -1.0 -1.0 -1.0 0.0536 0.3444
1.2502 6.0 558 1.0684 0.0381 0.1233 0.0015 0.0979 0.0184 -1.0 0.0593 0.2889 0.3556 0.44 0.3059 -1.0 -1.0 -1.0 0.0381 0.3556
1.2502 7.0 651 1.0613 0.0252 0.0884 0.0024 0.0409 0.0208 -1.0 0.0407 0.2593 0.3407 0.35 0.3353 -1.0 -1.0 -1.0 0.0252 0.3407
1.2502 8.0 744 0.9724 0.0602 0.2019 0.008 0.1258 0.034 -1.0 0.1407 0.3704 0.4296 0.5 0.3882 -1.0 -1.0 -1.0 0.0602 0.4296
1.2502 9.0 837 1.0170 0.055 0.1548 0.0077 0.0627 0.0556 -1.0 0.0667 0.3889 0.3889 0.41 0.3765 -1.0 -1.0 -1.0 0.055 0.3889
1.2502 10.0 930 1.1431 0.0324 0.1301 0.0025 0.0532 0.0281 -1.0 0.063 0.2741 0.3037 0.32 0.2941 -1.0 -1.0 -1.0 0.0324 0.3037
0.8688 11.0 1023 0.9432 0.0511 0.1614 0.0152 0.1059 0.0394 -1.0 0.0333 0.3407 0.4481 0.49 0.4235 -1.0 -1.0 -1.0 0.0511 0.4481
0.8688 12.0 1116 0.9395 0.0658 0.1824 0.0115 0.1559 0.0409 -1.0 0.0815 0.3741 0.4333 0.49 0.4 -1.0 -1.0 -1.0 0.0658 0.4333
0.8688 13.0 1209 1.0367 0.0554 0.1926 0.0138 0.123 0.0377 -1.0 0.1074 0.3 0.3667 0.47 0.3059 -1.0 -1.0 -1.0 0.0554 0.3667
0.8688 14.0 1302 1.0167 0.0459 0.1409 0.0077 0.0799 0.0452 -1.0 0.0444 0.3 0.4 0.46 0.3647 -1.0 -1.0 -1.0 0.0459 0.4
0.8688 15.0 1395 1.0687 0.0403 0.126 0.005 0.0494 0.0472 -1.0 0.0778 0.2741 0.3444 0.48 0.2647 -1.0 -1.0 -1.0 0.0403 0.3444
0.8688 16.0 1488 0.9219 0.1075 0.2652 0.0661 0.233 0.0563 -1.0 0.1185 0.3741 0.4889 0.58 0.4353 -1.0 -1.0 -1.0 0.1075 0.4889
0.7025 17.0 1581 0.9246 0.0706 0.2286 0.0293 0.2273 0.0339 -1.0 0.1111 0.3778 0.4444 0.58 0.3647 -1.0 -1.0 -1.0 0.0706 0.4444
0.7025 18.0 1674 0.9737 0.0638 0.1656 0.0298 0.1258 0.0524 -1.0 0.0778 0.4111 0.463 0.55 0.4118 -1.0 -1.0 -1.0 0.0638 0.463
0.7025 19.0 1767 0.9821 0.0687 0.2266 0.0189 0.1069 0.0594 -1.0 0.0444 0.3704 0.4259 0.46 0.4059 -1.0 -1.0 -1.0 0.0687 0.4259
0.7025 20.0 1860 0.9198 0.0966 0.3077 0.0152 0.212 0.058 -1.0 0.1519 0.3704 0.4778 0.59 0.4118 -1.0 -1.0 -1.0 0.0966 0.4778
0.7025 21.0 1953 0.9468 0.087 0.2415 0.0305 0.1982 0.0708 -1.0 0.163 0.3778 0.4667 0.51 0.4412 -1.0 -1.0 -1.0 0.087 0.4667
0.5994 22.0 2046 0.8969 0.1079 0.2414 0.0639 0.2735 0.0765 -1.0 0.1444 0.4296 0.5407 0.64 0.4824 -1.0 -1.0 -1.0 0.1079 0.5407
0.5994 23.0 2139 0.9532 0.0974 0.233 0.0462 0.249 0.0641 -1.0 0.163 0.3963 0.5111 0.62 0.4471 -1.0 -1.0 -1.0 0.0974 0.5111
0.5994 24.0 2232 0.9191 0.104 0.2925 0.0506 0.2068 0.0671 -1.0 0.1148 0.4 0.4926 0.62 0.4176 -1.0 -1.0 -1.0 0.104 0.4926
0.5994 25.0 2325 0.8977 0.1095 0.3031 0.0364 0.1874 0.0843 -1.0 0.1148 0.437 0.5296 0.63 0.4706 -1.0 -1.0 -1.0 0.1095 0.5296
0.5994 26.0 2418 0.9240 0.0999 0.2688 0.0423 0.1689 0.0872 -1.0 0.0963 0.4222 0.5185 0.59 0.4765 -1.0 -1.0 -1.0 0.0999 0.5185
0.526 27.0 2511 0.9139 0.1091 0.2691 0.038 0.2088 0.0872 -1.0 0.1259 0.4222 0.5185 0.6 0.4706 -1.0 -1.0 -1.0 0.1091 0.5185
0.526 28.0 2604 0.9143 0.1058 0.2494 0.0547 0.2091 0.0827 -1.0 0.1037 0.4296 0.5296 0.62 0.4765 -1.0 -1.0 -1.0 0.1058 0.5296
0.526 29.0 2697 0.9166 0.1103 0.2704 0.037 0.2046 0.0867 -1.0 0.1074 0.4259 0.5259 0.61 0.4765 -1.0 -1.0 -1.0 0.1103 0.5259
0.526 30.0 2790 0.9194 0.1092 0.2705 0.0385 0.2009 0.0867 -1.0 0.1037 0.4296 0.5222 0.61 0.4706 -1.0 -1.0 -1.0 0.1092 0.5222

Framework versions

  • Transformers 4.42.4
  • Pytorch 2.3.1+cu121
  • Datasets 2.20.0
  • Tokenizers 0.19.1