File size: 12,682 Bytes
a6e6635 ada438d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 |
2022-02-06 16:36:43,588 INFO MainThread:2903924 [wandb_setup.py:_flush():71] setting env: {'project': 'wav2vec2', 'entity': 'NbAiLab'}
2022-02-06 16:36:43,588 INFO MainThread:2903924 [wandb_setup.py:_flush():71] setting login settings: {}
2022-02-06 16:36:43,589 INFO MainThread:2903924 [wandb_init.py:_log_setup():371] Logging user logs to /mnt/lv_ai_1_dante/javierr/wav2vec2-xls-r-300m-npsc-bokmaal/wandb/run-20220206_163643-2773pjij/logs/debug.log
2022-02-06 16:36:43,589 INFO MainThread:2903924 [wandb_init.py:_log_setup():372] Logging internal logs to /mnt/lv_ai_1_dante/javierr/wav2vec2-xls-r-300m-npsc-bokmaal/wandb/run-20220206_163643-2773pjij/logs/debug-internal.log
2022-02-06 16:36:43,589 INFO MainThread:2903924 [wandb_init.py:init():404] calling init triggers
2022-02-06 16:36:43,589 INFO MainThread:2903924 [wandb_init.py:init():409] wandb.init called with sweep_config: {}
config: {}
2022-02-06 16:36:43,589 INFO MainThread:2903924 [wandb_init.py:init():460] starting backend
2022-02-06 16:36:43,589 INFO MainThread:2903924 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
2022-02-06 16:36:43,666 INFO MainThread:2903924 [backend.py:ensure_launched():216] starting backend process...
2022-02-06 16:36:43,760 INFO MainThread:2903924 [backend.py:ensure_launched():221] started backend process with pid: 2905110
2022-02-06 16:36:43,762 INFO MainThread:2903924 [wandb_init.py:init():469] backend started and connected
2022-02-06 16:36:43,770 INFO MainThread:2903924 [wandb_init.py:init():533] updated telemetry
2022-02-06 16:36:43,966 INFO MainThread:2903924 [wandb_init.py:init():563] communicating current version
2022-02-06 16:36:44,575 INFO MainThread:2903924 [wandb_init.py:init():568] got version response upgrade_message: "wandb version 0.12.10 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
2022-02-06 16:36:44,576 INFO MainThread:2903924 [wandb_init.py:init():578] communicating run to backend with 30 second timeout
2022-02-06 16:36:44,832 INFO MainThread:2903924 [wandb_init.py:init():606] starting run threads in backend
2022-02-06 16:36:46,297 INFO MainThread:2903924 [wandb_run.py:_console_start():1810] atexit reg
2022-02-06 16:36:46,297 INFO MainThread:2903924 [wandb_run.py:_redirect():1684] redirect: SettingsConsole.REDIRECT
2022-02-06 16:36:46,298 INFO MainThread:2903924 [wandb_run.py:_redirect():1689] Redirecting console.
2022-02-06 16:36:46,301 INFO MainThread:2903924 [wandb_run.py:_redirect():1745] Redirects installed.
2022-02-06 16:36:46,301 INFO MainThread:2903924 [wandb_init.py:init():633] run started, returning control to user process
2022-02-06 16:36:46,304 INFO MainThread:2903924 [wandb_run.py:_config_callback():956] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 31, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-300m', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'model_type': 'wav2vec2', 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.047, 'attention_dropout': 0.094, 'activation_dropout': 0.055, 'feat_proj_dropout': 0.04, 'final_dropout': 0.0, 'layerdrop': 0.041, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 34, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.082, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.25, 'mask_feature_length': 64, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'mean', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0001, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 2000, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Feb06_16-29-04_dante', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 100, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['tensorboard', 'wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'NbAiLab/wav2vec2-xls-r-300m-npsc-bokmaal', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'train_batch_size': 16, 'eval_batch_size': 16}
2022-02-06 16:36:46,310 INFO MainThread:2903924 [wandb_watch.py:watch():43] Watching
2022-02-07 09:41:34,032 INFO MainThread:2903924 [wandb_run.py:_atexit_cleanup():1780] got exitcode: 255
2022-02-07 09:41:34,053 INFO MainThread:2903924 [wandb_run.py:_restore():1752] restore
2022-02-07 09:41:36,648 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 1
other_count: 1
}
pusher_stats {
uploaded_bytes: 35316
total_bytes: 35316
}
2022-02-07 09:41:36,756 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 1
other_count: 1
}
pusher_stats {
uploaded_bytes: 35316
total_bytes: 35316
}
2022-02-07 09:41:37,540 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 1
other_count: 1
}
pusher_stats {
uploaded_bytes: 35316
total_bytes: 35316
}
2022-02-07 09:41:37,867 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
other_count: 1
}
pusher_stats {
uploaded_bytes: 35316
total_bytes: 10964440
}
2022-02-07 09:41:37,969 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
other_count: 1
}
pusher_stats {
uploaded_bytes: 35316
total_bytes: 10964440
}
2022-02-07 09:41:38,072 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
other_count: 1
}
pusher_stats {
uploaded_bytes: 35316
total_bytes: 10964440
}
2022-02-07 09:41:38,174 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
other_count: 1
}
pusher_stats {
uploaded_bytes: 35316
total_bytes: 10964440
}
2022-02-07 09:41:38,276 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
other_count: 1
}
pusher_stats {
uploaded_bytes: 1295154
total_bytes: 10964440
}
2022-02-07 09:41:38,379 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
other_count: 1
}
pusher_stats {
uploaded_bytes: 6765163
total_bytes: 10964440
}
2022-02-07 09:41:38,481 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
other_count: 1
}
pusher_stats {
uploaded_bytes: 10964440
total_bytes: 10964440
}
2022-02-07 09:41:38,583 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
other_count: 1
}
pusher_stats {
uploaded_bytes: 10964440
total_bytes: 10964440
}
2022-02-07 09:41:38,685 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
other_count: 1
}
pusher_stats {
uploaded_bytes: 10964440
total_bytes: 10964440
}
2022-02-07 09:41:38,787 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
other_count: 1
}
pusher_stats {
uploaded_bytes: 10964440
total_bytes: 10964440
}
2022-02-07 09:41:38,890 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
other_count: 1
}
pusher_stats {
uploaded_bytes: 10964440
total_bytes: 10964440
}
2022-02-07 09:41:38,992 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
other_count: 1
}
pusher_stats {
uploaded_bytes: 10964440
total_bytes: 10964440
}
2022-02-07 09:41:39,094 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
other_count: 1
}
pusher_stats {
uploaded_bytes: 10964440
total_bytes: 10964440
}
2022-02-07 09:41:40,110 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
other_count: 1
}
pusher_stats {
uploaded_bytes: 10964440
total_bytes: 10964440
}
2022-02-07 09:41:40,661 INFO MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: done: true
exit_result {
}
file_counts {
wandb_count: 5
other_count: 1
}
pusher_stats {
uploaded_bytes: 10964440
total_bytes: 10964440
}
local_info {
}
2022-02-07 09:41:42,062 INFO MainThread:2903924 [wandb_run.py:_append_history():2130] rendering history
2022-02-07 09:41:42,064 INFO MainThread:2903924 [wandb_run.py:_append_summary():2085] rendering summary
2022-02-07 09:41:42,070 INFO MainThread:2903924 [wandb_run.py:_append_files():2180] logging synced files
|