runs
This model is a fine-tuned version of on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 24.0950
- Accuracy: 0.0013
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 16
- eval_batch_size: 48
- seed: 444
- gradient_accumulation_steps: 3
- total_train_batch_size: 48
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine_with_restarts
- lr_scheduler_warmup_ratio: 0.3
- training_steps: 100000
- mixed_precision_training: Native AMP
Training results
Training Loss | Epoch | Step | Validation Loss | Accuracy |
---|---|---|---|---|
8.2359 | 6.04 | 1000 | 8.2170 | 0.0070 |
7.7137 | 12.07 | 2000 | 7.7007 | 0.0064 |
6.5277 | 18.11 | 3000 | 6.5254 | 0.0000 |
6.0375 | 24.14 | 4000 | 6.0532 | 0.0000 |
5.6908 | 30.18 | 5000 | 5.7100 | 0.0001 |
5.4294 | 36.22 | 6000 | 5.4758 | 0.0002 |
5.2161 | 42.25 | 7000 | 5.2891 | 0.0006 |
5.0151 | 48.29 | 8000 | 5.1152 | 0.0021 |
4.8349 | 54.33 | 9000 | 4.9847 | 0.0020 |
4.6358 | 60.36 | 10000 | 4.8754 | 0.0022 |
4.4326 | 66.4 | 11000 | 4.7809 | 0.0021 |
4.2632 | 72.43 | 12000 | 4.7416 | 0.0017 |
4.0415 | 78.47 | 13000 | 4.7503 | 0.0016 |
3.8196 | 84.51 | 14000 | 4.8472 | 0.0014 |
3.6207 | 90.54 | 15000 | 5.0215 | 0.0014 |
3.3163 | 96.58 | 16000 | 5.2939 | 0.0014 |
3.0377 | 102.62 | 17000 | 5.6685 | 0.0014 |
2.7272 | 108.65 | 18000 | 6.1649 | 0.0013 |
2.4319 | 114.69 | 19000 | 6.7556 | 0.0013 |
2.1647 | 120.72 | 20000 | 7.3951 | 0.0013 |
1.9001 | 126.76 | 21000 | 8.0823 | 0.0013 |
1.6708 | 132.8 | 22000 | 8.8230 | 0.0013 |
1.4762 | 138.83 | 23000 | 9.5335 | 0.0013 |
1.2833 | 144.87 | 24000 | 10.1973 | 0.0013 |
1.1451 | 150.91 | 25000 | 10.8213 | 0.0013 |
1.0251 | 156.94 | 26000 | 11.4402 | 0.0013 |
0.9164 | 162.98 | 27000 | 11.9995 | 0.0013 |
0.8174 | 169.01 | 28000 | 12.5680 | 0.0013 |
0.6862 | 175.05 | 29000 | 13.0050 | 0.0013 |
0.5738 | 181.09 | 30000 | 13.4692 | 0.0013 |
0.4524 | 187.12 | 31000 | 13.9220 | 0.0013 |
0.4252 | 193.16 | 32000 | 14.3340 | 0.0013 |
0.3952 | 199.2 | 33000 | 14.7961 | 0.0013 |
0.3684 | 205.23 | 34000 | 15.2421 | 0.0013 |
0.3338 | 211.27 | 35000 | 15.6433 | 0.0013 |
0.307 | 217.3 | 36000 | 16.0182 | 0.0013 |
0.2951 | 223.34 | 37000 | 16.3087 | 0.0013 |
0.28 | 229.38 | 38000 | 16.6556 | 0.0013 |
0.2688 | 235.41 | 39000 | 16.9303 | 0.0013 |
0.2582 | 241.45 | 40000 | 17.2209 | 0.0013 |
0.238 | 247.48 | 41000 | 17.5311 | 0.0013 |
0.2261 | 253.52 | 42000 | 17.7731 | 0.0013 |
0.21 | 259.56 | 43000 | 18.0205 | 0.0013 |
0.2073 | 265.59 | 44000 | 18.2693 | 0.0013 |
0.1976 | 271.63 | 45000 | 18.4634 | 0.0013 |
0.1865 | 277.67 | 46000 | 18.7215 | 0.0012 |
0.1769 | 283.7 | 47000 | 18.9467 | 0.0013 |
0.1649 | 289.74 | 48000 | 19.1423 | 0.0013 |
0.1517 | 295.77 | 49000 | 19.3638 | 0.0013 |
0.1491 | 301.81 | 50000 | 19.5879 | 0.0013 |
0.1387 | 307.85 | 51000 | 19.7823 | 0.0013 |
0.1332 | 313.88 | 52000 | 19.9663 | 0.0013 |
0.1256 | 319.92 | 53000 | 20.1907 | 0.0013 |
0.1154 | 325.96 | 54000 | 20.3939 | 0.0013 |
0.1091 | 331.99 | 55000 | 20.5926 | 0.0013 |
0.0928 | 338.03 | 56000 | 20.8044 | 0.0013 |
0.0812 | 344.06 | 57000 | 20.9873 | 0.0013 |
0.0677 | 350.1 | 58000 | 21.1931 | 0.0013 |
0.0609 | 356.14 | 59000 | 21.3650 | 0.0013 |
0.058 | 362.17 | 60000 | 21.5868 | 0.0013 |
0.0532 | 368.21 | 61000 | 21.7740 | 0.0013 |
0.0481 | 374.25 | 62000 | 21.9339 | 0.0013 |
0.0358 | 380.28 | 63000 | 22.1660 | 0.0012 |
0.0117 | 386.32 | 64000 | 22.4226 | 0.0013 |
0.0768 | 392.35 | 65000 | 22.2193 | 0.0013 |
0.0339 | 398.39 | 66000 | 22.3833 | 0.0013 |
0.0191 | 404.43 | 67000 | 22.5927 | 0.0013 |
0.0493 | 410.46 | 68000 | 22.6069 | 0.0013 |
0.0115 | 416.5 | 69000 | 22.8652 | 0.0012 |
0.0111 | 422.54 | 70000 | 22.9982 | 0.0012 |
0.1182 | 428.57 | 71000 | 22.6628 | 0.0013 |
0.0118 | 434.61 | 72000 | 22.9036 | 0.0013 |
0.0111 | 440.64 | 73000 | 23.0692 | 0.0013 |
0.011 | 446.68 | 74000 | 23.1857 | 0.0013 |
0.0386 | 452.72 | 75000 | 22.9263 | 0.0013 |
0.0109 | 458.75 | 76000 | 23.1548 | 0.0013 |
0.0109 | 464.79 | 77000 | 23.2761 | 0.0012 |
0.0108 | 470.82 | 78000 | 23.3763 | 0.0013 |
0.0131 | 476.86 | 79000 | 23.2048 | 0.0013 |
0.0108 | 482.9 | 80000 | 23.3772 | 0.0013 |
0.0106 | 488.93 | 81000 | 23.4733 | 0.0013 |
0.0106 | 494.97 | 82000 | 23.5654 | 0.0013 |
0.0242 | 501.01 | 83000 | 23.5459 | 0.0013 |
0.0104 | 507.04 | 84000 | 23.5695 | 0.0013 |
0.01 | 513.08 | 85000 | 23.6659 | 0.0013 |
0.0098 | 519.11 | 86000 | 23.7337 | 0.0013 |
0.0097 | 525.15 | 87000 | 23.7961 | 0.0013 |
0.0097 | 531.19 | 88000 | 23.8573 | 0.0013 |
0.0097 | 537.22 | 89000 | 23.9052 | 0.0013 |
0.0097 | 543.26 | 90000 | 23.9524 | 0.0013 |
0.0096 | 549.3 | 91000 | 23.9823 | 0.0013 |
0.0096 | 555.33 | 92000 | 24.0084 | 0.0013 |
0.0095 | 561.37 | 93000 | 24.0364 | 0.0013 |
0.0095 | 567.4 | 94000 | 24.0545 | 0.0013 |
0.0094 | 573.44 | 95000 | 24.0701 | 0.0013 |
0.0094 | 579.48 | 96000 | 24.0826 | 0.0013 |
0.0093 | 585.51 | 97000 | 24.0898 | 0.0013 |
0.0093 | 591.55 | 98000 | 24.0935 | 0.0013 |
0.0093 | 597.59 | 99000 | 24.0944 | 0.0013 |
0.0092 | 603.62 | 100000 | 24.0950 | 0.0013 |
Framework versions
- Transformers 4.37.2
- Pytorch 2.2.0+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
- Downloads last month
- 9
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.