0%| | 100/503920 [03:55<127:33:26, 1.10it/s] 0%| | 199/503920 [07:55<136:12:40, 1.03it/s] 0%| | 299/503920 [11:55<140:30:23, 1.00s/it] 0%| | 398/503920 [15:47<145:22:16, 1.04s/it] 0%|▏ | 500/503920 [19:41<127:22:04, 1.10it/s]The following columns in the evaluation set don't have a corresponding argument in `Wav2Vec2ForCTC.forward` and have been ignored: input_length. If input_length are not expected by `Wav2Vec2ForCTC.forward`, you can safely ignore this message. ***** Running Evaluation ***** Num examples = 41040 Batch size = 12 {'loss': 2.9263, 'learning_rate': 4.9000000000000005e-06, 'epoch': 0.04} File "/mnt/lv_ai_1_dante/ml/models/wav2vec2-1b-npsc-nst-bokmaal/run_speech_recognition_ctc.py", line 819, in █████████████████████████████████████████▏ | 3297/3420 [43:23<01:41, 1.21it/s] main() File "/mnt/lv_ai_1_dante/ml/models/wav2vec2-1b-npsc-nst-bokmaal/run_speech_recognition_ctc.py", line 770, in main train_result = trainer.train(resume_from_checkpoint=checkpoint) File "/mnt/lv_ai_1_dante/ml/rolvb/venv/lib/python3.9/site-packages/transformers/trainer.py", line 1497, in train self._maybe_log_save_evaluate(tr_loss, model, trial, epoch, ignore_keys_for_eval) File "/mnt/lv_ai_1_dante/ml/rolvb/venv/lib/python3.9/site-packages/transformers/trainer.py", line 1624, in _maybe_log_save_evaluate metrics = self.evaluate(ignore_keys=ignore_keys_for_eval) File "/mnt/lv_ai_1_dante/ml/rolvb/venv/lib/python3.9/site-packages/transformers/trainer.py", line 2284, in evaluate output = eval_loop( File "/mnt/lv_ai_1_dante/ml/rolvb/venv/lib/python3.9/site-packages/transformers/trainer.py", line 2448, in evaluation_loop for step, inputs in enumerate(dataloader): File "/mnt/lv_ai_1_dante/ml/rolvb/venv/lib/python3.9/site-packages/torch/utils/data/dataloader.py", line 530, in __next__ data = self._next_data() File "/mnt/lv_ai_1_dante/ml/rolvb/venv/lib/python3.9/site-packages/torch/utils/data/dataloader.py", line 570, in _next_data data = self._dataset_fetcher.fetch(index) # may raise StopIteration File "/mnt/lv_ai_1_dante/ml/rolvb/venv/lib/python3.9/site-packages/torch/utils/data/_utils/fetch.py", line 52, in fetch return self.collate_fn(data) File "/mnt/lv_ai_1_dante/ml/models/wav2vec2-1b-npsc-nst-bokmaal/run_speech_recognition_ctc.py", line 281, in __call__ batch = self.processor.pad( File "/mnt/lv_ai_1_dante/ml/rolvb/venv/lib/python3.9/site-packages/transformers/models/wav2vec2/processing_wav2vec2.py", line 82, in pad return self.current_processor.pad(*args, **kwargs) File "/mnt/lv_ai_1_dante/ml/rolvb/venv/lib/python3.9/site-packages/transformers/feature_extraction_sequence_utils.py", line 178, in pad processed_features[key] = [to_numpy(v) for v in value] File "/mnt/lv_ai_1_dante/ml/rolvb/venv/lib/python3.9/site-packages/transformers/feature_extraction_sequence_utils.py", line 178, in processed_features[key] = [to_numpy(v) for v in value] File "/mnt/lv_ai_1_dante/ml/rolvb/venv/lib/python3.9/site-packages/transformers/utils/generic.py", line 135, in to_numpy return np.array(obj) KeyboardInterrupt