Training in progress, step 200
Browse files- .ipynb_checkpoints/requirements-checkpoint.txt +8 -0
- .ipynb_checkpoints/run-checkpoint.sh +14 -12
- config.json +4 -4
- pytorch_model.bin +1 -1
- requirements.txt +6 -1
- run.sh +14 -12
- runs/Feb05_08-15-45_job-0a778896-a7e2-46e9-bcf5-016f91f242cf/1644049006.003602/events.out.tfevents.1644049006.job-0a778896-a7e2-46e9-bcf5-016f91f242cf.23457.1 +3 -0
- runs/Feb05_08-15-45_job-0a778896-a7e2-46e9-bcf5-016f91f242cf/events.out.tfevents.1644049005.job-0a778896-a7e2-46e9-bcf5-016f91f242cf.23457.0 +3 -0
- special_tokens_map.json +1 -1
- tokenizer_config.json +1 -1
- training_args.bin +1 -1
.ipynb_checkpoints/requirements-checkpoint.txt
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
unidecode
|
2 |
+
tensorboard
|
3 |
+
torch
|
4 |
+
torchaudio
|
5 |
+
jiwer~=2.3.0
|
6 |
+
soundfile~=0.10.3
|
7 |
+
transformers~=4.16.2
|
8 |
+
datasets~=1.18.3
|
.ipynb_checkpoints/run-checkpoint.sh
CHANGED
@@ -1,29 +1,31 @@
|
|
1 |
-
python xls-r-uzbek-cv8/run_speech_recognition_ctc.py \
|
2 |
--dataset_name="mozilla-foundation/common_voice_8_0" \
|
3 |
--model_name_or_path="facebook/wav2vec2-xls-r-300m" \
|
4 |
--dataset_config_name="uz" \
|
5 |
-
--output_dir="
|
|
|
|
|
6 |
--overwrite_output_dir \
|
7 |
-
--num_train_epochs="
|
8 |
-
--per_device_train_batch_size="
|
9 |
--per_device_eval_batch_size="8" \
|
10 |
--gradient_accumulation_steps="4" \
|
11 |
--learning_rate="3e-5" \
|
12 |
-
--warmup_steps="
|
13 |
--length_column_name="input_length" \
|
14 |
--evaluation_strategy="steps" \
|
15 |
--text_column_name="sentence" \
|
16 |
-
--eval_metrics="
|
17 |
-
--save_steps="
|
18 |
-
--eval_steps="
|
19 |
--logging_steps="100" \
|
20 |
-
--min_duration_in_seconds="0.
|
21 |
-
--layerdrop="0.
|
22 |
--activation_dropout="0.1" \
|
23 |
--save_total_limit="3" \
|
24 |
--freeze_feature_encoder \
|
25 |
-
--feat_proj_dropout="0.
|
26 |
-
--mask_time_prob="0.
|
27 |
--mask_time_length="10" \
|
28 |
--mask_feature_prob="0.25" \
|
29 |
--mask_feature_length="64" \
|
|
|
1 |
+
python ~/xls-r-uzbek-cv8/run_speech_recognition_ctc.py \
|
2 |
--dataset_name="mozilla-foundation/common_voice_8_0" \
|
3 |
--model_name_or_path="facebook/wav2vec2-xls-r-300m" \
|
4 |
--dataset_config_name="uz" \
|
5 |
+
--output_dir="~/xls-r-uzbek-cv8" \
|
6 |
+
--train_split_name="train[:30%]" \
|
7 |
+
--eval_split_name="validation[50%:]" \
|
8 |
--overwrite_output_dir \
|
9 |
+
--num_train_epochs="100" \
|
10 |
+
--per_device_train_batch_size="32" \
|
11 |
--per_device_eval_batch_size="8" \
|
12 |
--gradient_accumulation_steps="4" \
|
13 |
--learning_rate="3e-5" \
|
14 |
+
--warmup_steps="500" \
|
15 |
--length_column_name="input_length" \
|
16 |
--evaluation_strategy="steps" \
|
17 |
--text_column_name="sentence" \
|
18 |
+
--eval_metrics="cer" \
|
19 |
+
--save_steps="200" \
|
20 |
+
--eval_steps="200" \
|
21 |
--logging_steps="100" \
|
22 |
+
--min_duration_in_seconds="0.2" \
|
23 |
+
--layerdrop="0.01" \
|
24 |
--activation_dropout="0.1" \
|
25 |
--save_total_limit="3" \
|
26 |
--freeze_feature_encoder \
|
27 |
+
--feat_proj_dropout="0.01" \
|
28 |
+
--mask_time_prob="0.50" \
|
29 |
--mask_time_length="10" \
|
30 |
--mask_feature_prob="0.25" \
|
31 |
--mask_feature_length="64" \
|
config.json
CHANGED
@@ -49,7 +49,7 @@
|
|
49 |
"feat_extract_activation": "gelu",
|
50 |
"feat_extract_dropout": 0.0,
|
51 |
"feat_extract_norm": "layer",
|
52 |
-
"feat_proj_dropout": 0.
|
53 |
"feat_quantizer_dropout": 0.0,
|
54 |
"final_dropout": 0.0,
|
55 |
"hidden_act": "gelu",
|
@@ -58,13 +58,13 @@
|
|
58 |
"initializer_range": 0.02,
|
59 |
"intermediate_size": 4096,
|
60 |
"layer_norm_eps": 1e-05,
|
61 |
-
"layerdrop": 0.
|
62 |
"mask_feature_length": 64,
|
63 |
"mask_feature_min_masks": 0,
|
64 |
"mask_feature_prob": 0.25,
|
65 |
"mask_time_length": 10,
|
66 |
"mask_time_min_masks": 2,
|
67 |
-
"mask_time_prob": 0.
|
68 |
"model_type": "wav2vec2",
|
69 |
"num_adapter_layers": 3,
|
70 |
"num_attention_heads": 16,
|
@@ -100,7 +100,7 @@
|
|
100 |
1
|
101 |
],
|
102 |
"torch_dtype": "float32",
|
103 |
-
"transformers_version": "4.
|
104 |
"use_weighted_layer_sum": false,
|
105 |
"vocab_size": 33,
|
106 |
"xvector_output_dim": 512,
|
|
|
49 |
"feat_extract_activation": "gelu",
|
50 |
"feat_extract_dropout": 0.0,
|
51 |
"feat_extract_norm": "layer",
|
52 |
+
"feat_proj_dropout": 0.01,
|
53 |
"feat_quantizer_dropout": 0.0,
|
54 |
"final_dropout": 0.0,
|
55 |
"hidden_act": "gelu",
|
|
|
58 |
"initializer_range": 0.02,
|
59 |
"intermediate_size": 4096,
|
60 |
"layer_norm_eps": 1e-05,
|
61 |
+
"layerdrop": 0.01,
|
62 |
"mask_feature_length": 64,
|
63 |
"mask_feature_min_masks": 0,
|
64 |
"mask_feature_prob": 0.25,
|
65 |
"mask_time_length": 10,
|
66 |
"mask_time_min_masks": 2,
|
67 |
+
"mask_time_prob": 0.5,
|
68 |
"model_type": "wav2vec2",
|
69 |
"num_adapter_layers": 3,
|
70 |
"num_attention_heads": 16,
|
|
|
100 |
1
|
101 |
],
|
102 |
"torch_dtype": "float32",
|
103 |
+
"transformers_version": "4.16.2",
|
104 |
"use_weighted_layer_sum": false,
|
105 |
"vocab_size": 33,
|
106 |
"xvector_output_dim": 512,
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1262058993
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3f107f42a91fd5d39621347b43afb0c87ef3258c4bc185ec165f2870c082b35c
|
3 |
size 1262058993
|
requirements.txt
CHANGED
@@ -1,3 +1,8 @@
|
|
1 |
unidecode
|
2 |
tensorboard
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
1 |
unidecode
|
2 |
tensorboard
|
3 |
+
torch
|
4 |
+
torchaudio
|
5 |
+
jiwer~=2.3.0
|
6 |
+
soundfile~=0.10.3
|
7 |
+
transformers~=4.16.2
|
8 |
+
datasets~=1.18.3
|
run.sh
CHANGED
@@ -1,29 +1,31 @@
|
|
1 |
-
python xls-r-uzbek-cv8/run_speech_recognition_ctc.py \
|
2 |
--dataset_name="mozilla-foundation/common_voice_8_0" \
|
3 |
--model_name_or_path="facebook/wav2vec2-xls-r-300m" \
|
4 |
--dataset_config_name="uz" \
|
5 |
-
--output_dir="
|
|
|
|
|
6 |
--overwrite_output_dir \
|
7 |
-
--num_train_epochs="
|
8 |
-
--per_device_train_batch_size="
|
9 |
--per_device_eval_batch_size="8" \
|
10 |
--gradient_accumulation_steps="4" \
|
11 |
--learning_rate="3e-5" \
|
12 |
-
--warmup_steps="
|
13 |
--length_column_name="input_length" \
|
14 |
--evaluation_strategy="steps" \
|
15 |
--text_column_name="sentence" \
|
16 |
-
--eval_metrics="
|
17 |
-
--save_steps="
|
18 |
-
--eval_steps="
|
19 |
--logging_steps="100" \
|
20 |
-
--min_duration_in_seconds="0.
|
21 |
-
--layerdrop="0.
|
22 |
--activation_dropout="0.1" \
|
23 |
--save_total_limit="3" \
|
24 |
--freeze_feature_encoder \
|
25 |
-
--feat_proj_dropout="0.
|
26 |
-
--mask_time_prob="0.
|
27 |
--mask_time_length="10" \
|
28 |
--mask_feature_prob="0.25" \
|
29 |
--mask_feature_length="64" \
|
|
|
1 |
+
python ~/xls-r-uzbek-cv8/run_speech_recognition_ctc.py \
|
2 |
--dataset_name="mozilla-foundation/common_voice_8_0" \
|
3 |
--model_name_or_path="facebook/wav2vec2-xls-r-300m" \
|
4 |
--dataset_config_name="uz" \
|
5 |
+
--output_dir="~/xls-r-uzbek-cv8" \
|
6 |
+
--train_split_name="train[:30%]" \
|
7 |
+
--eval_split_name="validation[50%:]" \
|
8 |
--overwrite_output_dir \
|
9 |
+
--num_train_epochs="100" \
|
10 |
+
--per_device_train_batch_size="32" \
|
11 |
--per_device_eval_batch_size="8" \
|
12 |
--gradient_accumulation_steps="4" \
|
13 |
--learning_rate="3e-5" \
|
14 |
+
--warmup_steps="500" \
|
15 |
--length_column_name="input_length" \
|
16 |
--evaluation_strategy="steps" \
|
17 |
--text_column_name="sentence" \
|
18 |
+
--eval_metrics="cer" \
|
19 |
+
--save_steps="200" \
|
20 |
+
--eval_steps="200" \
|
21 |
--logging_steps="100" \
|
22 |
+
--min_duration_in_seconds="0.2" \
|
23 |
+
--layerdrop="0.01" \
|
24 |
--activation_dropout="0.1" \
|
25 |
--save_total_limit="3" \
|
26 |
--freeze_feature_encoder \
|
27 |
+
--feat_proj_dropout="0.01" \
|
28 |
+
--mask_time_prob="0.50" \
|
29 |
--mask_time_length="10" \
|
30 |
--mask_feature_prob="0.25" \
|
31 |
--mask_feature_length="64" \
|
runs/Feb05_08-15-45_job-0a778896-a7e2-46e9-bcf5-016f91f242cf/1644049006.003602/events.out.tfevents.1644049006.job-0a778896-a7e2-46e9-bcf5-016f91f242cf.23457.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:84f165c6e40f5a1e87d21f624a9b4475d58842f270044fb50735b8f5a9d00101
|
3 |
+
size 4826
|
runs/Feb05_08-15-45_job-0a778896-a7e2-46e9-bcf5-016f91f242cf/events.out.tfevents.1644049005.job-0a778896-a7e2-46e9-bcf5-016f91f242cf.23457.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:314b3baf73fa8b9becdd35cfeecc2a0a871edb0fda354e78bfa86e375dcc81fd
|
3 |
+
size 5404
|
special_tokens_map.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
|
|
|
1 |
+
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
|
tokenizer_config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "
|
|
|
1 |
+
{"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "/workspace/xls-r-uzbek-cv8", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3055
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7a661f5e8e620b98c6c94d0bb8c2ddd7376f2412096a9ce2ff8088711cea618b
|
3 |
size 3055
|