salbatarni's picture
End of training
144b378 verified
metadata
license: apache-2.0
base_model: google-bert/bert-base-cased
tags:
  - generated_from_trainer
model-index:
  - name: bert_baseline_language_task3_fold3
    results: []

bert_baseline_language_task3_fold3

This model is a fine-tuned version of google-bert/bert-base-cased on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.3641
  • Qwk: 0.6679
  • Mse: 0.3649

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 5

Training results

Training Loss Epoch Step Validation Loss Qwk Mse
No log 0.0308 2 0.8306 0.1251 0.8302
No log 0.0615 4 0.6714 0.0129 0.6712
No log 0.0923 6 0.5734 0.0099 0.5733
No log 0.1231 8 0.5391 0.0 0.5393
No log 0.1538 10 0.5546 0.0 0.5548
No log 0.1846 12 0.5052 0.0 0.5056
No log 0.2154 14 0.4823 0.0099 0.4826
No log 0.2462 16 0.4737 0.0785 0.4741
No log 0.2769 18 0.4572 0.2068 0.4577
No log 0.3077 20 0.4337 0.3354 0.4342
No log 0.3385 22 0.4440 0.4915 0.4447
No log 0.3692 24 0.4478 0.5378 0.4486
No log 0.4 26 0.4114 0.5344 0.4120
No log 0.4308 28 0.3993 0.4289 0.3999
No log 0.4615 30 0.4059 0.3404 0.4062
No log 0.4923 32 0.3825 0.4644 0.3830
No log 0.5231 34 0.5125 0.6055 0.5136
No log 0.5538 36 0.5777 0.6026 0.5791
No log 0.5846 38 0.5047 0.6024 0.5059
No log 0.6154 40 0.3807 0.5499 0.3814
No log 0.6462 42 0.3562 0.5243 0.3567
No log 0.6769 44 0.3538 0.5252 0.3543
No log 0.7077 46 0.3781 0.5832 0.3789
No log 0.7385 48 0.4093 0.6070 0.4102
No log 0.7692 50 0.4201 0.6186 0.4211
No log 0.8 52 0.3967 0.6138 0.3976
No log 0.8308 54 0.3672 0.6082 0.3679
No log 0.8615 56 0.3377 0.5356 0.3381
No log 0.8923 58 0.3798 0.3861 0.3797
No log 0.9231 60 0.3843 0.3592 0.3841
No log 0.9538 62 0.3314 0.4222 0.3316
No log 0.9846 64 0.3160 0.5263 0.3165
No log 1.0154 66 0.3763 0.6104 0.3773
No log 1.0462 68 0.4321 0.6350 0.4332
No log 1.0769 70 0.3996 0.6316 0.4007
No log 1.1077 72 0.3290 0.5336 0.3297
No log 1.1385 74 0.3185 0.4143 0.3188
No log 1.1692 76 0.3147 0.4391 0.3150
No log 1.2 78 0.3078 0.5279 0.3083
No log 1.2308 80 0.3069 0.5593 0.3075
No log 1.2615 82 0.3144 0.5810 0.3151
No log 1.2923 84 0.3365 0.6023 0.3373
No log 1.3231 86 0.3249 0.5561 0.3255
No log 1.3538 88 0.3339 0.4817 0.3342
No log 1.3846 90 0.3238 0.5266 0.3242
No log 1.4154 92 0.3559 0.6035 0.3569
No log 1.4462 94 0.4087 0.6601 0.4100
No log 1.4769 96 0.4345 0.6539 0.4359
No log 1.5077 98 0.4370 0.6555 0.4384
No log 1.5385 100 0.3598 0.6397 0.3609
No log 1.5692 102 0.3039 0.5583 0.3044
No log 1.6 104 0.2976 0.5335 0.2980
No log 1.6308 106 0.3030 0.5666 0.3036
No log 1.6615 108 0.3040 0.5691 0.3046
No log 1.6923 110 0.3130 0.5853 0.3137
No log 1.7231 112 0.3382 0.6151 0.3391
No log 1.7538 114 0.3543 0.6348 0.3554
No log 1.7846 116 0.3649 0.6422 0.3660
No log 1.8154 118 0.3698 0.6487 0.3709
No log 1.8462 120 0.3528 0.6447 0.3538
No log 1.8769 122 0.3751 0.6637 0.3762
No log 1.9077 124 0.3971 0.6610 0.3983
No log 1.9385 126 0.3565 0.6653 0.3575
No log 1.9692 128 0.3034 0.5766 0.3040
No log 2.0 130 0.2948 0.5480 0.2954
No log 2.0308 132 0.3070 0.5827 0.3077
No log 2.0615 134 0.3637 0.6566 0.3647
No log 2.0923 136 0.4103 0.6631 0.4115
No log 2.1231 138 0.4045 0.6643 0.4057
No log 2.1538 140 0.3552 0.6667 0.3561
No log 2.1846 142 0.3109 0.5513 0.3113
No log 2.2154 144 0.3670 0.4582 0.3669
No log 2.2462 146 0.3712 0.4662 0.3711
No log 2.2769 148 0.3292 0.5312 0.3295
No log 2.3077 150 0.3480 0.6310 0.3488
No log 2.3385 152 0.3888 0.6526 0.3900
No log 2.3692 154 0.4500 0.6605 0.4514
No log 2.4 156 0.4250 0.6620 0.4263
No log 2.4308 158 0.3710 0.6552 0.3720
No log 2.4615 160 0.3132 0.6115 0.3138
No log 2.4923 162 0.2975 0.5737 0.2980
No log 2.5231 164 0.2938 0.5846 0.2944
No log 2.5538 166 0.3015 0.6116 0.3022
No log 2.5846 168 0.3224 0.6264 0.3231
No log 2.6154 170 0.3834 0.6689 0.3845
No log 2.6462 172 0.4118 0.6639 0.4130
No log 2.6769 174 0.3792 0.6745 0.3802
No log 2.7077 176 0.3231 0.6318 0.3239
No log 2.7385 178 0.2894 0.5780 0.2898
No log 2.7692 180 0.2889 0.5501 0.2892
No log 2.8 182 0.3062 0.6314 0.3068
No log 2.8308 184 0.3842 0.6697 0.3852
No log 2.8615 186 0.4438 0.6625 0.4451
No log 2.8923 188 0.4239 0.6741 0.4251
No log 2.9231 190 0.3771 0.6622 0.3782
No log 2.9538 192 0.3535 0.6459 0.3544
No log 2.9846 194 0.3286 0.6182 0.3293
No log 3.0154 196 0.3348 0.6233 0.3356
No log 3.0462 198 0.3786 0.6738 0.3796
No log 3.0769 200 0.4293 0.6788 0.4306
No log 3.1077 202 0.4096 0.6720 0.4108
No log 3.1385 204 0.3490 0.6591 0.3498
No log 3.1692 206 0.3197 0.6380 0.3203
No log 3.2 208 0.3003 0.5838 0.3007
No log 3.2308 210 0.2955 0.5548 0.2959
No log 3.2615 212 0.2989 0.5682 0.2994
No log 3.2923 214 0.3209 0.6273 0.3215
No log 3.3231 216 0.3816 0.6754 0.3825
No log 3.3538 218 0.4800 0.6771 0.4813
No log 3.3846 220 0.5092 0.6806 0.5106
No log 3.4154 222 0.4528 0.6903 0.4539
No log 3.4462 224 0.3572 0.6482 0.3579
No log 3.4769 226 0.3218 0.6212 0.3223
No log 3.5077 228 0.3252 0.6340 0.3257
No log 3.5385 230 0.3529 0.6686 0.3537
No log 3.5692 232 0.4101 0.6885 0.4112
No log 3.6 234 0.4086 0.6840 0.4096
No log 3.6308 236 0.3599 0.6712 0.3607
No log 3.6615 238 0.3240 0.6440 0.3246
No log 3.6923 240 0.3211 0.6373 0.3217
No log 3.7231 242 0.3433 0.6565 0.3440
No log 3.7538 244 0.3885 0.6755 0.3895
No log 3.7846 246 0.4180 0.6726 0.4191
No log 3.8154 248 0.4314 0.6774 0.4326
No log 3.8462 250 0.4021 0.6798 0.4031
No log 3.8769 252 0.3549 0.6614 0.3557
No log 3.9077 254 0.3336 0.6568 0.3343
No log 3.9385 256 0.3341 0.6591 0.3348
No log 3.9692 258 0.3482 0.6618 0.3489
No log 4.0 260 0.3594 0.6665 0.3602
No log 4.0308 262 0.3616 0.6613 0.3624
No log 4.0615 264 0.3774 0.6586 0.3782
No log 4.0923 266 0.3964 0.6596 0.3974
No log 4.1231 268 0.3869 0.6655 0.3878
No log 4.1538 270 0.3862 0.6623 0.3871
No log 4.1846 272 0.3721 0.6554 0.3729
No log 4.2154 274 0.3661 0.6583 0.3669
No log 4.2462 276 0.3606 0.6623 0.3614
No log 4.2769 278 0.3679 0.6590 0.3687
No log 4.3077 280 0.3805 0.6565 0.3814
No log 4.3385 282 0.3757 0.6586 0.3766
No log 4.3692 284 0.3656 0.6621 0.3664
No log 4.4 286 0.3581 0.6565 0.3589
No log 4.4308 288 0.3514 0.6586 0.3521
No log 4.4615 290 0.3543 0.6602 0.3550
No log 4.4923 292 0.3532 0.6565 0.3539
No log 4.5231 294 0.3648 0.6648 0.3656
No log 4.5538 296 0.3731 0.6710 0.3739
No log 4.5846 298 0.3856 0.6712 0.3865
No log 4.6154 300 0.4048 0.6726 0.4057
No log 4.6462 302 0.4102 0.6772 0.4112
No log 4.6769 304 0.4119 0.6803 0.4129
No log 4.7077 306 0.4111 0.6811 0.4121
No log 4.7385 308 0.4052 0.6735 0.4062
No log 4.7692 310 0.3954 0.6753 0.3964
No log 4.8 312 0.3905 0.6730 0.3914
No log 4.8308 314 0.3836 0.6670 0.3845
No log 4.8615 316 0.3768 0.6640 0.3777
No log 4.8923 318 0.3712 0.6697 0.3721
No log 4.9231 320 0.3676 0.6679 0.3684
No log 4.9538 322 0.3652 0.6679 0.3660
No log 4.9846 324 0.3641 0.6679 0.3649

Framework versions

  • Transformers 4.42.3
  • Pytorch 2.1.2
  • Datasets 2.20.0
  • Tokenizers 0.19.1