File size: 14,443 Bytes
5a2b856 5c97d18 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 |
2022-05-23 11:51:45,597 INFO MainThread:1049120 [wandb_setup.py:_flush():75] Loading settings from /home/rolvb/.config/wandb/settings
2022-05-23 11:51:45,597 INFO MainThread:1049120 [wandb_setup.py:_flush():75] Loading settings from /mnt/lv_ai_1_dante/ml/models/wav2vec2-1b-npsc-nst-bokmaal/wandb/settings
2022-05-23 11:51:45,597 INFO MainThread:1049120 [wandb_setup.py:_flush():75] Loading settings from environment variables: {'project': 'wav2vec2', 'entity': 'NbAiLab'}
2022-05-23 11:51:45,597 INFO MainThread:1049120 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_ctc.py', 'program': '/mnt/lv_ai_1_dante/ml/models/wav2vec2-1b-npsc-nst-bokmaal/run_speech_recognition_ctc.py'}
2022-05-23 11:51:45,597 INFO MainThread:1049120 [wandb_init.py:_log_setup():437] Logging user logs to /mnt/lv_ai_1_dante/ml/models/wav2vec2-1b-npsc-nst-bokmaal/wandb/run-20220523_115145-3dybzmyz/logs/debug.log
2022-05-23 11:51:45,597 INFO MainThread:1049120 [wandb_init.py:_log_setup():438] Logging internal logs to /mnt/lv_ai_1_dante/ml/models/wav2vec2-1b-npsc-nst-bokmaal/wandb/run-20220523_115145-3dybzmyz/logs/debug-internal.log
2022-05-23 11:51:45,598 INFO MainThread:1049120 [wandb_init.py:init():471] calling init triggers
2022-05-23 11:51:45,598 INFO MainThread:1049120 [wandb_init.py:init():474] wandb.init called with sweep_config: {}
config: {}
2022-05-23 11:51:45,598 INFO MainThread:1049120 [wandb_init.py:init():524] starting backend
2022-05-23 11:51:45,598 INFO MainThread:1049120 [backend.py:_multiprocessing_setup():97] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
2022-05-23 11:51:45,689 INFO MainThread:1049120 [backend.py:ensure_launched():217] starting backend process...
2022-05-23 11:51:45,775 INFO MainThread:1049120 [backend.py:ensure_launched():222] started backend process with pid: 1049978
2022-05-23 11:51:45,776 INFO MainThread:1049120 [wandb_init.py:init():533] backend started and connected
2022-05-23 11:51:45,787 INFO MainThread:1049120 [wandb_init.py:init():597] updated telemetry
2022-05-23 11:51:45,996 INFO MainThread:1049120 [wandb_init.py:init():628] communicating run to backend with 30 second timeout
2022-05-23 11:51:46,532 INFO MainThread:1049120 [wandb_run.py:_on_init():1923] communicating current version
2022-05-23 11:51:46,660 INFO MainThread:1049120 [wandb_run.py:_on_init():1927] got version response upgrade_message: "wandb version 0.12.16 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
2022-05-23 11:51:46,660 INFO MainThread:1049120 [wandb_init.py:init():659] starting run threads in backend
2022-05-23 11:51:46,714 INFO MainThread:1049120 [wandb_run.py:_console_start():1897] atexit reg
2022-05-23 11:51:46,715 INFO MainThread:1049120 [wandb_run.py:_redirect():1770] redirect: SettingsConsole.REDIRECT
2022-05-23 11:51:46,716 INFO MainThread:1049120 [wandb_run.py:_redirect():1775] Redirecting console.
2022-05-23 11:51:46,718 INFO MainThread:1049120 [wandb_run.py:_redirect():1831] Redirects installed.
2022-05-23 11:51:46,719 INFO MainThread:1049120 [wandb_init.py:init():684] run started, returning control to user process
2022-05-23 11:51:46,741 INFO MainThread:1049120 [wandb_run.py:_config_callback():1131] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 31, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-1b', 'transformers_version': '4.18.0', 'feat_extract_dropout': 0.0, 'model_type': 'wav2vec2', 'num_feat_extract_layers': 7, 'hidden_size': 1280, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 48, 'intermediate_size': 5120, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.047, 'attention_dropout': 0.094, 'activation_dropout': 0.055, 'feat_proj_dropout': 0.04, 'final_dropout': 0.0, 'layerdrop': 0.041, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 34, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.082, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.25, 'mask_feature_length': 64, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 1024, 'proj_codevector_dim': 1024, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'mean', 'ctc_zero_infinity': True, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1280, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 12, 'per_device_eval_batch_size': 12, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 'None', 'eval_delay': 0, 'learning_rate': 2e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 40.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 2000, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/May23_11-33-45_dante', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 100, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'data_seed': 'None', 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'NbAiLab/wav2vec2-1b-npsc-nst-bokmaal', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'train_batch_size': 12, 'eval_batch_size': 12}
2022-05-23 11:51:46,744 INFO MainThread:1049120 [wandb_watch.py:watch():47] Watching
2022-06-16 03:42:56,632 INFO MainThread:1049120 [wandb_run.py:_atexit_cleanup():1866] got exitcode: 1
2022-06-16 03:42:56,637 INFO MainThread:1049120 [wandb_run.py:_restore():1838] restore
2022-06-16 03:42:58,963 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 1
}
pusher_stats {
uploaded_bytes: 2176
total_bytes: 2176
}
2022-06-16 03:42:59,142 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 1
}
pusher_stats {
uploaded_bytes: 2176
total_bytes: 2176
}
2022-06-16 03:43:00,491 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 1
}
pusher_stats {
uploaded_bytes: 2176
total_bytes: 2176
}
2022-06-16 03:43:01,085 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 2176
total_bytes: 9696920
}
2022-06-16 03:43:01,187 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 2176
total_bytes: 9696920
}
2022-06-16 03:43:01,290 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 2176
total_bytes: 9696920
}
2022-06-16 03:43:01,392 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 2176
total_bytes: 9696920
}
2022-06-16 03:43:01,495 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 249250
total_bytes: 9696920
}
2022-06-16 03:43:01,597 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 1702809
total_bytes: 9696920
}
2022-06-16 03:43:01,700 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 6595559
total_bytes: 9696920
}
2022-06-16 03:43:01,802 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 9696920
total_bytes: 9696920
}
2022-06-16 03:43:01,904 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 9696920
total_bytes: 9696920
}
2022-06-16 03:43:02,007 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 9696920
total_bytes: 9696920
}
2022-06-16 03:43:02,109 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 9696920
total_bytes: 9696920
}
2022-06-16 03:43:02,211 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 9696920
total_bytes: 9696920
}
2022-06-16 03:43:02,314 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 9696920
total_bytes: 9696920
}
2022-06-16 03:43:02,416 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 9696920
total_bytes: 9696920
}
2022-06-16 03:43:02,518 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 9696920
total_bytes: 9696920
}
2022-06-16 03:43:02,621 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 9696920
total_bytes: 9696920
}
2022-06-16 03:43:02,723 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 9696920
total_bytes: 9696920
}
2022-06-16 03:43:02,825 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 9696920
total_bytes: 9696920
}
2022-06-16 03:43:02,928 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 9696920
total_bytes: 9696920
}
2022-06-16 03:43:03,030 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 9696920
total_bytes: 9696920
}
2022-06-16 03:43:03,132 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 9696920
total_bytes: 9696920
}
2022-06-16 03:43:04,949 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 9696920
total_bytes: 9696920
}
2022-06-16 03:43:05,548 INFO MainThread:1049120 [wandb_run.py:_on_finish():1995] got exit ret: done: true
exit_result {
}
file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 9696920
total_bytes: 9696920
}
local_info {
}
2022-06-16 03:43:06,750 INFO MainThread:1049120 [wandb_run.py:_footer_history_summary_info():3102] rendering history
2022-06-16 03:43:06,789 INFO MainThread:1049120 [wandb_run.py:_footer_history_summary_info():3134] rendering summary
2022-06-16 03:43:06,795 INFO MainThread:1049120 [wandb_run.py:_footer_sync_info():3057] logging synced files
|