jun-han's picture
Model save
33fb8b7 verified
|
raw
history blame
2.56 kB
metadata
library_name: transformers
license: apache-2.0
base_model: jun-han/Whisper-Small-architecture-change
tags:
  - generated_from_trainer
model-index:
  - name: Whisper-Small-architecture-change
    results: []

Whisper-Small-architecture-change

This model is a fine-tuned version of jun-han/Whisper-Small-architecture-change on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4281
  • Cer: 24.9352

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-05
  • train_batch_size: 16
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 3750
  • training_steps: 60000
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Cer Validation Loss
3.0206 1.0 3750 101.4703 2.9448
2.6624 2.0 7500 99.0366 2.7587
1.4476 3.0 11250 80.5695 1.4832
0.7219 4.0 15000 49.1456 0.8063
0.5387 5.0 18750 41.2985 0.6620
0.3824 6.0 22500 39.1312 0.6240
0.3445 7.0 26250 37.5137 0.5984
0.2819 8.0 30000 36.6006 0.5823
0.538 9.0 33750 0.5156 30.9133
0.3792 10.0 37500 0.4562 27.3205
0.2974 11.0 41250 0.4408 26.4521
0.2281 12.0 45000 0.4225 25.7254
0.1644 13.0 48750 0.4207 25.1123
0.1208 14.0 52500 0.4241 25.0098
0.0877 15.0 56250 0.4232 24.9781
0.0652 16.0 60000 0.4281 24.9352

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0
  • Datasets 2.21.0
  • Tokenizers 0.19.1