Rolv-Arild commited on
Commit
39df5ba
1 Parent(s): e008db8

Model save

Browse files
special_tokens_map.json CHANGED
@@ -1 +1 @@
1
- {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0a4c62ed4fe00460fa88aedac2e0ba57edf7f3671f2df41c5d284a51e8b6f867
3
  size 3055
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f682e2dfbdc6c55979388a4ba7cad6af122c433f2313ff3a5c222c5de1261c9
3
  size 3055
wandb/debug-internal.log CHANGED
@@ -1 +1 @@
1
- run-20220622_135628-3qxrqni1/logs/debug-internal.log
 
1
+ run-20220718_100633-rrwx85kq/logs/debug-internal.log
wandb/debug.log CHANGED
@@ -1 +1 @@
1
- run-20220622_135628-3qxrqni1/logs/debug.log
 
1
+ run-20220718_100633-rrwx85kq/logs/debug.log
wandb/latest-run CHANGED
@@ -1 +1 @@
1
- run-20220622_135628-3qxrqni1
 
1
+ run-20220718_100633-rrwx85kq
wandb/run-20220622_135628-3qxrqni1/files/config.yaml CHANGED
@@ -9441,6 +9441,14 @@ _wandb:
9441
  - 51
9442
  - 53
9443
  - 55
 
 
 
 
 
 
 
 
9444
  3:
9445
  - 1
9446
  - 7
 
9441
  - 51
9442
  - 53
9443
  - 55
9444
+ 2:
9445
+ - 1
9446
+ - 5
9447
+ - 11
9448
+ - 49
9449
+ - 51
9450
+ - 53
9451
+ - 55
9452
  3:
9453
  - 1
9454
  - 7
wandb/run-20220622_135628-3qxrqni1/files/output.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc31ec4aca3a51573de3b02f7f0cb1085f66186d55cbed7b76d2ed38fa065426
3
- size 262148635
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3eeb8f1b4218c7b8603c31ce64b37ad03c8f67453e12b17bb5011a9dcce7217
3
+ size 262353822
wandb/run-20220622_135628-3qxrqni1/files/wandb-summary.json CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/run-20220622_135628-3qxrqni1/logs/debug-internal.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b3715606cd059352650834b3f9f7a2765176411229cadf53219940d538292f35
3
- size 639277298
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2c062996592cc310a42bd05578afe3a3a4c1001b7749bf09dd659a3d30651b4
3
+ size 639474173
wandb/run-20220622_135628-3qxrqni1/logs/debug.log CHANGED
@@ -25,3 +25,325 @@ config: {}
25
  2022-06-22 13:56:30,375 INFO MainThread:383050 [wandb_init.py:init():684] run started, returning control to user process
26
  2022-06-22 13:56:30,397 INFO MainThread:383050 [wandb_run.py:_config_callback():1131] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 31, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-1b', 'transformers_version': '4.18.0', 'feat_extract_dropout': 0.0, 'model_type': 'wav2vec2', 'num_feat_extract_layers': 7, 'hidden_size': 1280, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 48, 'intermediate_size': 5120, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.047, 'attention_dropout': 0.094, 'activation_dropout': 0.055, 'feat_proj_dropout': 0.04, 'final_dropout': 0.0, 'layerdrop': 0.041, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 34, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.082, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.25, 'mask_feature_length': 64, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 1024, 'proj_codevector_dim': 1024, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'mean', 'ctc_zero_infinity': True, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1280, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': './', 'overwrite_output_dir': False, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 12, 'per_device_eval_batch_size': 12, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 'None', 'eval_delay': 0, 'learning_rate': 2e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 40.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 2000, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Jun22_13-51-42_dante', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 100, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'data_seed': 'None', 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'NbAiLab/wav2vec2-1b-npsc-nst-bokmaal', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'train_batch_size': 12, 'eval_batch_size': 12}
27
  2022-06-22 13:56:30,401 INFO MainThread:383050 [wandb_watch.py:watch():47] Watching
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
25
  2022-06-22 13:56:30,375 INFO MainThread:383050 [wandb_init.py:init():684] run started, returning control to user process
26
  2022-06-22 13:56:30,397 INFO MainThread:383050 [wandb_run.py:_config_callback():1131] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 31, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-1b', 'transformers_version': '4.18.0', 'feat_extract_dropout': 0.0, 'model_type': 'wav2vec2', 'num_feat_extract_layers': 7, 'hidden_size': 1280, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 48, 'intermediate_size': 5120, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.047, 'attention_dropout': 0.094, 'activation_dropout': 0.055, 'feat_proj_dropout': 0.04, 'final_dropout': 0.0, 'layerdrop': 0.041, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 34, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.082, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.25, 'mask_feature_length': 64, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 1024, 'proj_codevector_dim': 1024, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'mean', 'ctc_zero_infinity': True, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1280, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': './', 'overwrite_output_dir': False, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 12, 'per_device_eval_batch_size': 12, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 'None', 'eval_delay': 0, 'learning_rate': 2e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 40.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 2000, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Jun22_13-51-42_dante', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 100, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'data_seed': 'None', 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'NbAiLab/wav2vec2-1b-npsc-nst-bokmaal', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'train_batch_size': 12, 'eval_batch_size': 12}
27
  2022-06-22 13:56:30,401 INFO MainThread:383050 [wandb_watch.py:watch():47] Watching
28
+ 2022-07-14 18:44:11,351 INFO MainThread:383050 [wandb_run.py:_atexit_cleanup():1866] got exitcode: 1
29
+ 2022-07-14 18:44:11,354 INFO MainThread:383050 [wandb_run.py:_restore():1838] restore
30
+ 2022-07-14 18:44:13,690 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
31
+ wandb_count: 1
32
+ }
33
+ pusher_stats {
34
+ uploaded_bytes: 2142
35
+ total_bytes: 2142
36
+ }
37
+
38
+ 2022-07-14 18:44:13,842 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
39
+ wandb_count: 1
40
+ }
41
+ pusher_stats {
42
+ uploaded_bytes: 2142
43
+ total_bytes: 2142
44
+ }
45
+
46
+ 2022-07-14 18:44:15,980 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
47
+ wandb_count: 1
48
+ }
49
+ pusher_stats {
50
+ uploaded_bytes: 2142
51
+ total_bytes: 2142
52
+ }
53
+
54
+ 2022-07-14 18:44:16,636 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
55
+ wandb_count: 5
56
+ }
57
+ pusher_stats {
58
+ uploaded_bytes: 2142
59
+ total_bytes: 263999425
60
+ }
61
+
62
+ 2022-07-14 18:44:16,738 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
63
+ wandb_count: 5
64
+ }
65
+ pusher_stats {
66
+ uploaded_bytes: 2142
67
+ total_bytes: 263999425
68
+ }
69
+
70
+ 2022-07-14 18:44:16,840 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
71
+ wandb_count: 5
72
+ }
73
+ pusher_stats {
74
+ uploaded_bytes: 2142
75
+ total_bytes: 263999425
76
+ }
77
+
78
+ 2022-07-14 18:44:16,943 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
79
+ wandb_count: 5
80
+ }
81
+ pusher_stats {
82
+ uploaded_bytes: 2142
83
+ total_bytes: 263999425
84
+ }
85
+
86
+ 2022-07-14 18:44:17,045 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
87
+ wandb_count: 5
88
+ }
89
+ pusher_stats {
90
+ uploaded_bytes: 650624
91
+ total_bytes: 263999425
92
+ }
93
+
94
+ 2022-07-14 18:44:17,147 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
95
+ wandb_count: 5
96
+ }
97
+ pusher_stats {
98
+ uploaded_bytes: 7445539
99
+ total_bytes: 263999425
100
+ }
101
+
102
+ 2022-07-14 18:44:17,250 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
103
+ wandb_count: 5
104
+ }
105
+ pusher_stats {
106
+ uploaded_bytes: 20053027
107
+ total_bytes: 263999425
108
+ }
109
+
110
+ 2022-07-14 18:44:17,352 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
111
+ wandb_count: 5
112
+ }
113
+ pusher_stats {
114
+ uploaded_bytes: 32455715
115
+ total_bytes: 263999425
116
+ }
117
+
118
+ 2022-07-14 18:44:17,454 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
119
+ wandb_count: 5
120
+ }
121
+ pusher_stats {
122
+ uploaded_bytes: 45677603
123
+ total_bytes: 263999425
124
+ }
125
+
126
+ 2022-07-14 18:44:17,556 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
127
+ wandb_count: 5
128
+ }
129
+ pusher_stats {
130
+ uploaded_bytes: 58031139
131
+ total_bytes: 263999425
132
+ }
133
+
134
+ 2022-07-14 18:44:17,659 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
135
+ wandb_count: 5
136
+ }
137
+ pusher_stats {
138
+ uploaded_bytes: 70458403
139
+ total_bytes: 263999425
140
+ }
141
+
142
+ 2022-07-14 18:44:17,761 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
143
+ wandb_count: 5
144
+ }
145
+ pusher_stats {
146
+ uploaded_bytes: 83958819
147
+ total_bytes: 263999425
148
+ }
149
+
150
+ 2022-07-14 18:44:17,863 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
151
+ wandb_count: 5
152
+ }
153
+ pusher_stats {
154
+ uploaded_bytes: 96541731
155
+ total_bytes: 263999425
156
+ }
157
+
158
+ 2022-07-14 18:44:17,966 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
159
+ wandb_count: 5
160
+ }
161
+ pusher_stats {
162
+ uploaded_bytes: 109780003
163
+ total_bytes: 263999425
164
+ }
165
+
166
+ 2022-07-14 18:44:18,068 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
167
+ wandb_count: 5
168
+ }
169
+ pusher_stats {
170
+ uploaded_bytes: 120634403
171
+ total_bytes: 263999425
172
+ }
173
+
174
+ 2022-07-14 18:44:18,170 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
175
+ wandb_count: 5
176
+ }
177
+ pusher_stats {
178
+ uploaded_bytes: 134929443
179
+ total_bytes: 263999425
180
+ }
181
+
182
+ 2022-07-14 18:44:18,273 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
183
+ wandb_count: 5
184
+ }
185
+ pusher_stats {
186
+ uploaded_bytes: 146119715
187
+ total_bytes: 263999425
188
+ }
189
+
190
+ 2022-07-14 18:44:18,375 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
191
+ wandb_count: 5
192
+ }
193
+ pusher_stats {
194
+ uploaded_bytes: 156965923
195
+ total_bytes: 263999425
196
+ }
197
+
198
+ 2022-07-14 18:44:18,477 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
199
+ wandb_count: 5
200
+ }
201
+ pusher_stats {
202
+ uploaded_bytes: 170646563
203
+ total_bytes: 263999425
204
+ }
205
+
206
+ 2022-07-14 18:44:18,579 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
207
+ wandb_count: 5
208
+ }
209
+ pusher_stats {
210
+ uploaded_bytes: 183606307
211
+ total_bytes: 263999425
212
+ }
213
+
214
+ 2022-07-14 18:44:18,681 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
215
+ wandb_count: 5
216
+ }
217
+ pusher_stats {
218
+ uploaded_bytes: 193616931
219
+ total_bytes: 263999425
220
+ }
221
+
222
+ 2022-07-14 18:44:18,784 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
223
+ wandb_count: 5
224
+ }
225
+ pusher_stats {
226
+ uploaded_bytes: 206806051
227
+ total_bytes: 263999425
228
+ }
229
+
230
+ 2022-07-14 18:44:18,886 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
231
+ wandb_count: 5
232
+ }
233
+ pusher_stats {
234
+ uploaded_bytes: 220978211
235
+ total_bytes: 263999425
236
+ }
237
+
238
+ 2022-07-14 18:44:18,988 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
239
+ wandb_count: 5
240
+ }
241
+ pusher_stats {
242
+ uploaded_bytes: 230079523
243
+ total_bytes: 263999425
244
+ }
245
+
246
+ 2022-07-14 18:44:19,091 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
247
+ wandb_count: 5
248
+ }
249
+ pusher_stats {
250
+ uploaded_bytes: 244014115
251
+ total_bytes: 263999425
252
+ }
253
+
254
+ 2022-07-14 18:44:19,193 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
255
+ wandb_count: 5
256
+ }
257
+ pusher_stats {
258
+ uploaded_bytes: 256588835
259
+ total_bytes: 263999425
260
+ }
261
+
262
+ 2022-07-14 18:44:19,295 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
263
+ wandb_count: 5
264
+ }
265
+ pusher_stats {
266
+ uploaded_bytes: 263999425
267
+ total_bytes: 263999425
268
+ }
269
+
270
+ 2022-07-14 18:44:19,397 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
271
+ wandb_count: 5
272
+ }
273
+ pusher_stats {
274
+ uploaded_bytes: 263999425
275
+ total_bytes: 263999425
276
+ }
277
+
278
+ 2022-07-14 18:44:19,500 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
279
+ wandb_count: 5
280
+ }
281
+ pusher_stats {
282
+ uploaded_bytes: 263999425
283
+ total_bytes: 263999425
284
+ }
285
+
286
+ 2022-07-14 18:44:19,602 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
287
+ wandb_count: 5
288
+ }
289
+ pusher_stats {
290
+ uploaded_bytes: 263999425
291
+ total_bytes: 263999425
292
+ }
293
+
294
+ 2022-07-14 18:44:19,704 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
295
+ wandb_count: 5
296
+ }
297
+ pusher_stats {
298
+ uploaded_bytes: 263999425
299
+ total_bytes: 263999425
300
+ }
301
+
302
+ 2022-07-14 18:44:19,806 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
303
+ wandb_count: 5
304
+ }
305
+ pusher_stats {
306
+ uploaded_bytes: 263999425
307
+ total_bytes: 263999425
308
+ }
309
+
310
+ 2022-07-14 18:44:19,908 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
311
+ wandb_count: 5
312
+ }
313
+ pusher_stats {
314
+ uploaded_bytes: 263999425
315
+ total_bytes: 263999425
316
+ }
317
+
318
+ 2022-07-14 18:44:20,011 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
319
+ wandb_count: 5
320
+ }
321
+ pusher_stats {
322
+ uploaded_bytes: 263999425
323
+ total_bytes: 263999425
324
+ }
325
+
326
+ 2022-07-14 18:44:21,404 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: file_counts {
327
+ wandb_count: 5
328
+ }
329
+ pusher_stats {
330
+ uploaded_bytes: 263999425
331
+ total_bytes: 263999425
332
+ }
333
+
334
+ 2022-07-14 18:44:21,954 INFO MainThread:383050 [wandb_run.py:_on_finish():1995] got exit ret: done: true
335
+ exit_result {
336
+ }
337
+ file_counts {
338
+ wandb_count: 5
339
+ }
340
+ pusher_stats {
341
+ uploaded_bytes: 263999425
342
+ total_bytes: 263999425
343
+ }
344
+ local_info {
345
+ }
346
+
347
+ 2022-07-14 18:44:23,144 INFO MainThread:383050 [wandb_run.py:_footer_history_summary_info():3102] rendering history
348
+ 2022-07-14 18:44:23,183 INFO MainThread:383050 [wandb_run.py:_footer_history_summary_info():3134] rendering summary
349
+ 2022-07-14 18:44:23,190 INFO MainThread:383050 [wandb_run.py:_footer_sync_info():3057] logging synced files
wandb/run-20220622_135628-3qxrqni1/run-3qxrqni1.wandb CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:50f4e1b114a9b9a907b55811799fb713abd6253abf4cc01a3d4911320bc0b302
3
- size 3609828189
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92ddc96979af922dc9dc7e17e30f8218d1f7322deffe7f9c216837c914e7ee57
3
+ size 3611634765
wandb/run-20220718_100633-rrwx85kq/files/config.yaml ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20220718_100633-rrwx85kq/files/output.log ADDED
@@ -0,0 +1,160 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ 0%| | 0/499680 [00:00<?, ?it/s]
3
+
4
+
5
+
6
+
7
+
8
+
9
+
10
+
11
+
12
+
13
+
14
+
15
+
16
+
17
+
18
+
19
+
20
+
21
+
22
+
23
+
24
+
25
+
26
+
27
+
28
+
29
+
30
+
31
+
32
+
33
+
34
+
35
+
36
+
37
+
38
+
39
+
40
+
41
+
42
+
43
+
44
+
45
+
46
+
47
+
48
+
49
+
50
+
51
+
52
+
53
+
54
+
55
+
56
+
57
+
58
+
59
+
60
+
61
+
62
+
63
+
64
+
65
+
66
+
67
+
68
+
69
+
70
+
71
+
72
+
73
+
74
+
75
+
76
+
77
+
78
+
79
+
80
+
81
+
82
+
83
+
84
+ 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▉| 499600/499680 [2:32:23<04:12, 3.16s/it]
85
+
86
+
87
+
88
+
89
+
90
+
91
+
92
+
93
+
94
+
95
+
96
+
97
+
98
+
99
+
100
+
101
+
102
+
103
+
104
+
105
+
106
+
107
+
108
+
109
+
110
+
111
+
112
+
113
+
114
+
115
+
116
+
117
+
118
+
119
+
120
+
121
+
122
+
123
+
124
+
125
+
126
+
127
+
128
+
129
+
130
+
131
+
132
+
133
+
134
+
135
+
136
+
137
+
138
+
139
+
140
+
141
+
142
+
143
+
144
+
145
+
146
+
147
+
148
+
149
+
150
+ 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▉| 499677/499680 [2:35:07<00:03, 1.13s/it]
151
+ 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 499680/499680 [2:35:10<00:00, 1.06it/s]
152
+ Training completed. Do not forget to share your model on huggingface.co/models =)
153
+ 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 499680/499680 [2:35:10<00:00, 53.67it/s]
154
+ Saving model checkpoint to ./████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 499680/499680 [2:35:10<00:00, 1.06it/s]
155
+ Configuration saved in ./config.json
156
+ Model weights saved in ./pytorch_model.bin
157
+ Feature extractor saved in ./preprocessor_config.json
158
+ Saving model checkpoint to ./
159
+ Configuration saved in ./config.json
160
+ Model weights saved in ./pytorch_model.bin
wandb/run-20220718_100633-rrwx85kq/files/requirements.txt ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ aiohttp==3.8.1
2
+ aiosignal==1.2.0
3
+ appdirs==1.4.4
4
+ async-timeout==4.0.2
5
+ attrs==21.4.0
6
+ audioread==2.1.9
7
+ certifi==2021.10.8
8
+ cffi==1.15.0
9
+ charset-normalizer==2.0.12
10
+ click==8.1.2
11
+ datasets==2.1.0
12
+ decorator==5.1.1
13
+ dill==0.3.4
14
+ docker-pycreds==0.4.0
15
+ filelock==3.6.0
16
+ frozenlist==1.3.0
17
+ fsspec==2022.3.0
18
+ gitdb==4.0.9
19
+ gitpython==3.1.27
20
+ huggingface-hub==0.5.1
21
+ hypothesis==6.46.5
22
+ idna==3.3
23
+ jiwer==2.3.0
24
+ joblib==1.1.0
25
+ kenlm==0.0.0
26
+ librosa==0.9.1
27
+ llvmlite==0.38.0
28
+ multidict==6.0.2
29
+ multiprocess==0.70.12.2
30
+ numba==0.55.1
31
+ numpy==1.21.6
32
+ packaging==21.3
33
+ pandas==1.4.2
34
+ pathtools==0.1.2
35
+ pillow==9.1.0
36
+ pip==20.3.4
37
+ pkg-resources==0.0.0
38
+ pooch==1.6.0
39
+ promise==2.3
40
+ protobuf==3.20.1
41
+ psutil==5.9.0
42
+ pyarrow==7.0.0
43
+ pycparser==2.21
44
+ pyctcdecode==0.3.0
45
+ pygtrie==2.4.2
46
+ pyparsing==3.0.8
47
+ python-dateutil==2.8.2
48
+ python-levenshtein==0.12.2
49
+ pytz==2022.1
50
+ pyyaml==6.0
51
+ regex==2022.4.24
52
+ requests==2.27.1
53
+ resampy==0.2.2
54
+ responses==0.18.0
55
+ sacremoses==0.0.49
56
+ scikit-learn==1.0.2
57
+ scipy==1.8.0
58
+ sentry-sdk==1.5.10
59
+ setproctitle==1.2.3
60
+ setuptools==44.1.1
61
+ shortuuid==1.0.8
62
+ six==1.16.0
63
+ smmap==5.0.0
64
+ sortedcontainers==2.4.0
65
+ soundfile==0.10.3.post1
66
+ threadpoolctl==3.1.0
67
+ tokenizers==0.12.1
68
+ torch==1.11.0+cu113
69
+ torchaudio==0.11.0+cu113
70
+ torchvision==0.12.0+cu113
71
+ tqdm==4.64.0
72
+ transformers==4.18.0
73
+ typing-extensions==4.2.0
74
+ urllib3==1.26.9
75
+ wandb==0.12.15
76
+ xxhash==3.0.0
77
+ yarl==1.7.2
wandb/run-20220718_100633-rrwx85kq/files/wandb-metadata.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.13.0-40-generic-x86_64-with-glibc2.34",
3
+ "python": "3.9.7",
4
+ "heartbeatAt": "2022-07-18T08:06:36.267713",
5
+ "startedAt": "2022-07-18T08:06:33.716532",
6
+ "docker": null,
7
+ "cpu_count": 96,
8
+ "cuda": null,
9
+ "args": [
10
+ "--model_name_or_path=facebook/wav2vec2-xls-r-1b",
11
+ "--hub_model_id=NbAiLab/wav2vec2-1b-npsc-nst-bokmaal",
12
+ "--output_dir=./",
13
+ "--num_train_epochs=40",
14
+ "--per_device_train_batch_size=12",
15
+ "--per_device_eval_batch_size=12",
16
+ "--gradient_accumulation_steps=2",
17
+ "--learning_rate=2e-5",
18
+ "--warmup_steps=2000",
19
+ "--length_column_name=input_length",
20
+ "--evaluation_strategy=steps",
21
+ "--text_column_name=text",
22
+ "--save_steps=500",
23
+ "--eval_steps=500",
24
+ "--logging_steps=100",
25
+ "--layerdrop=0.041",
26
+ "--attention_dropout=0.094",
27
+ "--activation_dropout=0.055",
28
+ "--hidden_dropout=0.047",
29
+ "--save_total_limit=3",
30
+ "--freeze_feature_encoder",
31
+ "--feat_proj_dropout=0.04",
32
+ "--mask_time_prob=0.082",
33
+ "--mask_time_length=10",
34
+ "--mask_feature_prob=0.25",
35
+ "--mask_feature_length=64",
36
+ "--gradient_checkpointing",
37
+ "--min_duration_in_seconds=0.5",
38
+ "--max_duration_in_seconds=30.0",
39
+ "--use_auth_token",
40
+ "--seed=42",
41
+ "--fp16",
42
+ "--group_by_length",
43
+ "--do_train",
44
+ "--do_eval",
45
+ "--push_to_hub",
46
+ "--preprocessing_num_workers=32",
47
+ "--ctc_zero_infinity"
48
+ ],
49
+ "state": "running",
50
+ "program": "/mnt/lv_ai_1_dante/ml/models/wav2vec2-1b-npsc-nst-bokmaal/run_speech_recognition_ctc.py",
51
+ "codePath": "run_speech_recognition_ctc.py",
52
+ "git": {
53
+ "remote": "https://huggingface.co/NbAiLab/wav2vec2-1b-npsc-nst-bokmaal",
54
+ "commit": "e008db872d9d96db8509a7c68e084dc9329a6a7f"
55
+ },
56
+ "email": "[email protected]",
57
+ "root": "/mnt/lv_ai_1_dante/ml/models/wav2vec2-1b-npsc-nst-bokmaal",
58
+ "host": "dante",
59
+ "username": "rolvb",
60
+ "executable": "/mnt/lv_ai_1_dante/ml/rolvb/venv/bin/python"
61
+ }
wandb/run-20220718_100633-rrwx85kq/files/wandb-summary.json ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20220718_100633-rrwx85kq/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20220718_100633-rrwx85kq/logs/debug.log ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-07-18 10:06:33,727 INFO MainThread:4012631 [wandb_setup.py:_flush():75] Loading settings from /home/rolvb/.config/wandb/settings
2
+ 2022-07-18 10:06:33,728 INFO MainThread:4012631 [wandb_setup.py:_flush():75] Loading settings from /mnt/lv_ai_1_dante/ml/models/wav2vec2-1b-npsc-nst-bokmaal/wandb/settings
3
+ 2022-07-18 10:06:33,728 INFO MainThread:4012631 [wandb_setup.py:_flush():75] Loading settings from environment variables: {'project': 'wav2vec2', 'entity': 'NbAiLab'}
4
+ 2022-07-18 10:06:33,728 INFO MainThread:4012631 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_ctc.py', 'program': '/mnt/lv_ai_1_dante/ml/models/wav2vec2-1b-npsc-nst-bokmaal/run_speech_recognition_ctc.py'}
5
+ 2022-07-18 10:06:33,728 INFO MainThread:4012631 [wandb_init.py:_log_setup():437] Logging user logs to /mnt/lv_ai_1_dante/ml/models/wav2vec2-1b-npsc-nst-bokmaal/wandb/run-20220718_100633-rrwx85kq/logs/debug.log
6
+ 2022-07-18 10:06:33,728 INFO MainThread:4012631 [wandb_init.py:_log_setup():438] Logging internal logs to /mnt/lv_ai_1_dante/ml/models/wav2vec2-1b-npsc-nst-bokmaal/wandb/run-20220718_100633-rrwx85kq/logs/debug-internal.log
7
+ 2022-07-18 10:06:33,728 INFO MainThread:4012631 [wandb_init.py:init():471] calling init triggers
8
+ 2022-07-18 10:06:33,729 INFO MainThread:4012631 [wandb_init.py:init():474] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-07-18 10:06:33,729 INFO MainThread:4012631 [wandb_init.py:init():524] starting backend
11
+ 2022-07-18 10:06:33,729 INFO MainThread:4012631 [backend.py:_multiprocessing_setup():97] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-07-18 10:06:34,303 INFO MainThread:4012631 [backend.py:ensure_launched():217] starting backend process...
13
+ 2022-07-18 10:06:34,762 INFO MainThread:4012631 [backend.py:ensure_launched():222] started backend process with pid: 4016051
14
+ 2022-07-18 10:06:34,766 INFO MainThread:4012631 [wandb_init.py:init():533] backend started and connected
15
+ 2022-07-18 10:06:34,839 INFO MainThread:4012631 [wandb_init.py:init():597] updated telemetry
16
+ 2022-07-18 10:06:35,798 INFO MainThread:4012631 [wandb_init.py:init():628] communicating run to backend with 30 second timeout
17
+ 2022-07-18 10:06:36,095 INFO MainThread:4012631 [wandb_run.py:_on_init():1923] communicating current version
18
+ 2022-07-18 10:06:36,232 INFO MainThread:4012631 [wandb_run.py:_on_init():1927] got version response upgrade_message: "wandb version 0.12.21 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
19
+
20
+ 2022-07-18 10:06:36,232 INFO MainThread:4012631 [wandb_init.py:init():659] starting run threads in backend
21
+ 2022-07-18 10:06:36,306 INFO MainThread:4012631 [wandb_run.py:_console_start():1897] atexit reg
22
+ 2022-07-18 10:06:36,307 INFO MainThread:4012631 [wandb_run.py:_redirect():1770] redirect: SettingsConsole.REDIRECT
23
+ 2022-07-18 10:06:36,308 INFO MainThread:4012631 [wandb_run.py:_redirect():1775] Redirecting console.
24
+ 2022-07-18 10:06:36,310 INFO MainThread:4012631 [wandb_run.py:_redirect():1831] Redirects installed.
25
+ 2022-07-18 10:06:36,311 INFO MainThread:4012631 [wandb_init.py:init():684] run started, returning control to user process
26
+ 2022-07-18 10:06:36,335 INFO MainThread:4012631 [wandb_run.py:_config_callback():1131] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 31, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-1b', 'transformers_version': '4.18.0', 'feat_extract_dropout': 0.0, 'model_type': 'wav2vec2', 'num_feat_extract_layers': 7, 'hidden_size': 1280, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 48, 'intermediate_size': 5120, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.047, 'attention_dropout': 0.094, 'activation_dropout': 0.055, 'feat_proj_dropout': 0.04, 'final_dropout': 0.0, 'layerdrop': 0.041, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 34, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.082, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.25, 'mask_feature_length': 64, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 1024, 'proj_codevector_dim': 1024, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'mean', 'ctc_zero_infinity': True, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1280, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': './', 'overwrite_output_dir': False, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 12, 'per_device_eval_batch_size': 12, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 'None', 'eval_delay': 0, 'learning_rate': 2e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 40.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 2000, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Jul18_10-03-20_dante', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 100, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'data_seed': 'None', 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'NbAiLab/wav2vec2-1b-npsc-nst-bokmaal', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'train_batch_size': 12, 'eval_batch_size': 12}
27
+ 2022-07-18 10:06:36,340 INFO MainThread:4012631 [wandb_watch.py:watch():47] Watching
wandb/run-20220718_100633-rrwx85kq/run-rrwx85kq.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8ae4c79781156c1f4e6fe5fd9af0eeb3e5814f59461f53a643c040916dab655
3
+ size 2956154