File size: 12,682 Bytes
a6e6635
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ada438d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
2022-02-06 16:36:43,588 INFO    MainThread:2903924 [wandb_setup.py:_flush():71] setting env: {'project': 'wav2vec2', 'entity': 'NbAiLab'}
2022-02-06 16:36:43,588 INFO    MainThread:2903924 [wandb_setup.py:_flush():71] setting login settings: {}
2022-02-06 16:36:43,589 INFO    MainThread:2903924 [wandb_init.py:_log_setup():371] Logging user logs to /mnt/lv_ai_1_dante/javierr/wav2vec2-xls-r-300m-npsc-bokmaal/wandb/run-20220206_163643-2773pjij/logs/debug.log
2022-02-06 16:36:43,589 INFO    MainThread:2903924 [wandb_init.py:_log_setup():372] Logging internal logs to /mnt/lv_ai_1_dante/javierr/wav2vec2-xls-r-300m-npsc-bokmaal/wandb/run-20220206_163643-2773pjij/logs/debug-internal.log
2022-02-06 16:36:43,589 INFO    MainThread:2903924 [wandb_init.py:init():404] calling init triggers
2022-02-06 16:36:43,589 INFO    MainThread:2903924 [wandb_init.py:init():409] wandb.init called with sweep_config: {}
config: {}
2022-02-06 16:36:43,589 INFO    MainThread:2903924 [wandb_init.py:init():460] starting backend
2022-02-06 16:36:43,589 INFO    MainThread:2903924 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
2022-02-06 16:36:43,666 INFO    MainThread:2903924 [backend.py:ensure_launched():216] starting backend process...
2022-02-06 16:36:43,760 INFO    MainThread:2903924 [backend.py:ensure_launched():221] started backend process with pid: 2905110
2022-02-06 16:36:43,762 INFO    MainThread:2903924 [wandb_init.py:init():469] backend started and connected
2022-02-06 16:36:43,770 INFO    MainThread:2903924 [wandb_init.py:init():533] updated telemetry
2022-02-06 16:36:43,966 INFO    MainThread:2903924 [wandb_init.py:init():563] communicating current version
2022-02-06 16:36:44,575 INFO    MainThread:2903924 [wandb_init.py:init():568] got version response upgrade_message: "wandb version 0.12.10 is available!  To upgrade, please run:\n $ pip install wandb --upgrade"

2022-02-06 16:36:44,576 INFO    MainThread:2903924 [wandb_init.py:init():578] communicating run to backend with 30 second timeout
2022-02-06 16:36:44,832 INFO    MainThread:2903924 [wandb_init.py:init():606] starting run threads in backend
2022-02-06 16:36:46,297 INFO    MainThread:2903924 [wandb_run.py:_console_start():1810] atexit reg
2022-02-06 16:36:46,297 INFO    MainThread:2903924 [wandb_run.py:_redirect():1684] redirect: SettingsConsole.REDIRECT
2022-02-06 16:36:46,298 INFO    MainThread:2903924 [wandb_run.py:_redirect():1689] Redirecting console.
2022-02-06 16:36:46,301 INFO    MainThread:2903924 [wandb_run.py:_redirect():1745] Redirects installed.
2022-02-06 16:36:46,301 INFO    MainThread:2903924 [wandb_init.py:init():633] run started, returning control to user process
2022-02-06 16:36:46,304 INFO    MainThread:2903924 [wandb_run.py:_config_callback():956] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 31, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-300m', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'model_type': 'wav2vec2', 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.047, 'attention_dropout': 0.094, 'activation_dropout': 0.055, 'feat_proj_dropout': 0.04, 'final_dropout': 0.0, 'layerdrop': 0.041, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 34, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.082, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.25, 'mask_feature_length': 64, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'mean', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0001, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 2000, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Feb06_16-29-04_dante', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 100, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['tensorboard', 'wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'NbAiLab/wav2vec2-xls-r-300m-npsc-bokmaal', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'train_batch_size': 16, 'eval_batch_size': 16}
2022-02-06 16:36:46,310 INFO    MainThread:2903924 [wandb_watch.py:watch():43] Watching
2022-02-07 09:41:34,032 INFO    MainThread:2903924 [wandb_run.py:_atexit_cleanup():1780] got exitcode: 255
2022-02-07 09:41:34,053 INFO    MainThread:2903924 [wandb_run.py:_restore():1752] restore
2022-02-07 09:41:36,648 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 1
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 35316
  total_bytes: 35316
}

2022-02-07 09:41:36,756 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 1
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 35316
  total_bytes: 35316
}

2022-02-07 09:41:37,540 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 1
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 35316
  total_bytes: 35316
}

2022-02-07 09:41:37,867 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 35316
  total_bytes: 10964440
}

2022-02-07 09:41:37,969 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 35316
  total_bytes: 10964440
}

2022-02-07 09:41:38,072 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 35316
  total_bytes: 10964440
}

2022-02-07 09:41:38,174 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 35316
  total_bytes: 10964440
}

2022-02-07 09:41:38,276 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 1295154
  total_bytes: 10964440
}

2022-02-07 09:41:38,379 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 6765163
  total_bytes: 10964440
}

2022-02-07 09:41:38,481 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 10964440
  total_bytes: 10964440
}

2022-02-07 09:41:38,583 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 10964440
  total_bytes: 10964440
}

2022-02-07 09:41:38,685 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 10964440
  total_bytes: 10964440
}

2022-02-07 09:41:38,787 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 10964440
  total_bytes: 10964440
}

2022-02-07 09:41:38,890 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 10964440
  total_bytes: 10964440
}

2022-02-07 09:41:38,992 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 10964440
  total_bytes: 10964440
}

2022-02-07 09:41:39,094 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 10964440
  total_bytes: 10964440
}

2022-02-07 09:41:40,110 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 10964440
  total_bytes: 10964440
}

2022-02-07 09:41:40,661 INFO    MainThread:2903924 [wandb_run.py:_wait_for_finish():1912] got exit ret: done: true
exit_result {
}
file_counts {
  wandb_count: 5
  other_count: 1
}
pusher_stats {
  uploaded_bytes: 10964440
  total_bytes: 10964440
}
local_info {
}

2022-02-07 09:41:42,062 INFO    MainThread:2903924 [wandb_run.py:_append_history():2130] rendering history
2022-02-07 09:41:42,064 INFO    MainThread:2903924 [wandb_run.py:_append_summary():2085] rendering summary
2022-02-07 09:41:42,070 INFO    MainThread:2903924 [wandb_run.py:_append_files():2180] logging synced files