oliverguhr
commited on
Commit
•
422f0b1
1
Parent(s):
af7d143
final update
Browse files- README.md +114 -2
- all_results.json +14 -0
- eval_results.json +9 -0
- optimizer.pt +0 -3
- scheduler.pt → runs/Jun02_09-09-07_Niles/1654166101.6285756/events.out.tfevents.1654166101.Niles.467343.1 +2 -2
- rng_state.pth → runs/Jun02_09-09-07_Niles/events.out.tfevents.1654166101.Niles.467343.0 +2 -2
- scaler.pt → runs/Jun02_09-09-07_Niles/events.out.tfevents.1655605972.Niles.467343.2 +2 -2
- train_results.json +8 -0
- trainer_state.json +3615 -3
README.md
CHANGED
@@ -1,4 +1,116 @@
|
|
1 |
---
|
2 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
4 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
language:
|
3 |
+
- de
|
4 |
+
license: apache-2.0
|
5 |
+
tags:
|
6 |
+
- automatic-speech-recognition
|
7 |
+
- mozilla-foundation/common_voice_9_0
|
8 |
+
- generated_from_trainer
|
9 |
+
datasets:
|
10 |
+
- common_voice
|
11 |
+
model-index:
|
12 |
+
- name: wav2vec2-large-xlsr-53-german-cv9
|
13 |
+
results: []
|
14 |
---
|
15 |
+
|
16 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
17 |
+
should probably proofread and complete it, then remove this comment. -->
|
18 |
+
|
19 |
+
# wav2vec2-large-xlsr-53-german-cv9
|
20 |
+
|
21 |
+
This model is a fine-tuned version of [./facebook/wav2vec2-large-xlsr-53](https://huggingface.co/./facebook/wav2vec2-large-xlsr-53) on the MOZILLA-FOUNDATION/COMMON_VOICE_9_0 - DE dataset.
|
22 |
+
It achieves the following results on the evaluation set:
|
23 |
+
- Loss: 0.1093
|
24 |
+
- Wer: 0.0909
|
25 |
+
|
26 |
+
## Model description
|
27 |
+
|
28 |
+
More information needed
|
29 |
+
|
30 |
+
## Intended uses & limitations
|
31 |
+
|
32 |
+
More information needed
|
33 |
+
|
34 |
+
## Training and evaluation data
|
35 |
+
|
36 |
+
More information needed
|
37 |
+
|
38 |
+
## Training procedure
|
39 |
+
|
40 |
+
### Training hyperparameters
|
41 |
+
|
42 |
+
The following hyperparameters were used during training:
|
43 |
+
- learning_rate: 0.0001
|
44 |
+
- train_batch_size: 16
|
45 |
+
- eval_batch_size: 32
|
46 |
+
- seed: 42
|
47 |
+
- gradient_accumulation_steps: 8
|
48 |
+
- total_train_batch_size: 128
|
49 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
50 |
+
- lr_scheduler_type: linear
|
51 |
+
- lr_scheduler_warmup_ratio: 0.1
|
52 |
+
- num_epochs: 50.0
|
53 |
+
- mixed_precision_training: Native AMP
|
54 |
+
|
55 |
+
### Training results
|
56 |
+
|
57 |
+
| Training Loss | Epoch | Step | Validation Loss | Wer |
|
58 |
+
|:-------------:|:-----:|:------:|:---------------:|:------:|
|
59 |
+
| 0.4129 | 1.0 | 3557 | 0.3015 | 0.2499 |
|
60 |
+
| 0.2121 | 2.0 | 7114 | 0.1596 | 0.1567 |
|
61 |
+
| 0.1455 | 3.0 | 10671 | 0.1377 | 0.1354 |
|
62 |
+
| 0.1436 | 4.0 | 14228 | 0.1301 | 0.1282 |
|
63 |
+
| 0.1144 | 5.0 | 17785 | 0.1225 | 0.1245 |
|
64 |
+
| 0.1219 | 6.0 | 21342 | 0.1254 | 0.1208 |
|
65 |
+
| 0.104 | 7.0 | 24899 | 0.1198 | 0.1232 |
|
66 |
+
| 0.1016 | 8.0 | 28456 | 0.1149 | 0.1174 |
|
67 |
+
| 0.1093 | 9.0 | 32013 | 0.1186 | 0.1186 |
|
68 |
+
| 0.0858 | 10.0 | 35570 | 0.1182 | 0.1164 |
|
69 |
+
| 0.102 | 11.0 | 39127 | 0.1191 | 0.1186 |
|
70 |
+
| 0.0834 | 12.0 | 42684 | 0.1161 | 0.1096 |
|
71 |
+
| 0.0916 | 13.0 | 46241 | 0.1147 | 0.1107 |
|
72 |
+
| 0.0811 | 14.0 | 49798 | 0.1174 | 0.1136 |
|
73 |
+
| 0.0814 | 15.0 | 53355 | 0.1132 | 0.1114 |
|
74 |
+
| 0.0865 | 16.0 | 56912 | 0.1134 | 0.1097 |
|
75 |
+
| 0.0701 | 17.0 | 60469 | 0.1096 | 0.1054 |
|
76 |
+
| 0.0891 | 18.0 | 64026 | 0.1110 | 0.1076 |
|
77 |
+
| 0.071 | 19.0 | 67583 | 0.1141 | 0.1074 |
|
78 |
+
| 0.0726 | 20.0 | 71140 | 0.1094 | 0.1093 |
|
79 |
+
| 0.0647 | 21.0 | 74697 | 0.1088 | 0.1095 |
|
80 |
+
| 0.0643 | 22.0 | 78254 | 0.1105 | 0.1044 |
|
81 |
+
| 0.0764 | 23.0 | 81811 | 0.1072 | 0.1042 |
|
82 |
+
| 0.0605 | 24.0 | 85368 | 0.1095 | 0.1026 |
|
83 |
+
| 0.0722 | 25.0 | 88925 | 0.1144 | 0.1066 |
|
84 |
+
| 0.0597 | 26.0 | 92482 | 0.1087 | 0.1022 |
|
85 |
+
| 0.062 | 27.0 | 96039 | 0.1073 | 0.1027 |
|
86 |
+
| 0.0536 | 28.0 | 99596 | 0.1068 | 0.1027 |
|
87 |
+
| 0.0616 | 29.0 | 103153 | 0.1097 | 0.1037 |
|
88 |
+
| 0.0642 | 30.0 | 106710 | 0.1117 | 0.1020 |
|
89 |
+
| 0.0555 | 31.0 | 110267 | 0.1109 | 0.0990 |
|
90 |
+
| 0.0632 | 32.0 | 113824 | 0.1104 | 0.0977 |
|
91 |
+
| 0.0482 | 33.0 | 117381 | 0.1108 | 0.0958 |
|
92 |
+
| 0.0601 | 34.0 | 120938 | 0.1095 | 0.0957 |
|
93 |
+
| 0.0508 | 35.0 | 124495 | 0.1079 | 0.0973 |
|
94 |
+
| 0.0526 | 36.0 | 128052 | 0.1068 | 0.0967 |
|
95 |
+
| 0.0487 | 37.0 | 131609 | 0.1081 | 0.0966 |
|
96 |
+
| 0.0495 | 38.0 | 135166 | 0.1099 | 0.0956 |
|
97 |
+
| 0.0528 | 39.0 | 138723 | 0.1091 | 0.0923 |
|
98 |
+
| 0.0439 | 40.0 | 142280 | 0.1111 | 0.0928 |
|
99 |
+
| 0.0467 | 41.0 | 145837 | 0.1131 | 0.0943 |
|
100 |
+
| 0.0407 | 42.0 | 149394 | 0.1115 | 0.0944 |
|
101 |
+
| 0.046 | 43.0 | 152951 | 0.1106 | 0.0935 |
|
102 |
+
| 0.0447 | 44.0 | 156508 | 0.1083 | 0.0919 |
|
103 |
+
| 0.0434 | 45.0 | 160065 | 0.1093 | 0.0909 |
|
104 |
+
| 0.0472 | 46.0 | 163622 | 0.1092 | 0.0921 |
|
105 |
+
| 0.0414 | 47.0 | 167179 | 0.1106 | 0.0922 |
|
106 |
+
| 0.0501 | 48.0 | 170736 | 0.1094 | 0.0918 |
|
107 |
+
| 0.0388 | 49.0 | 174293 | 0.1099 | 0.0918 |
|
108 |
+
| 0.0428 | 50.0 | 177850 | 0.1103 | 0.0915 |
|
109 |
+
|
110 |
+
|
111 |
+
### Framework versions
|
112 |
+
|
113 |
+
- Transformers 4.19.0.dev0
|
114 |
+
- Pytorch 1.11.0+cu113
|
115 |
+
- Datasets 2.0.0
|
116 |
+
- Tokenizers 0.11.6
|
all_results.json
ADDED
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 50.0,
|
3 |
+
"eval_loss": 0.10925392806529999,
|
4 |
+
"eval_runtime": 635.7213,
|
5 |
+
"eval_samples": 16033,
|
6 |
+
"eval_samples_per_second": 25.22,
|
7 |
+
"eval_steps_per_second": 0.79,
|
8 |
+
"eval_wer": 0.09092025712063107,
|
9 |
+
"train_loss": 0.1478050081201038,
|
10 |
+
"train_runtime": 1439234.2576,
|
11 |
+
"train_samples": 455356,
|
12 |
+
"train_samples_per_second": 15.819,
|
13 |
+
"train_steps_per_second": 0.124
|
14 |
+
}
|
eval_results.json
ADDED
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 50.0,
|
3 |
+
"eval_loss": 0.10925392806529999,
|
4 |
+
"eval_runtime": 635.7213,
|
5 |
+
"eval_samples": 16033,
|
6 |
+
"eval_samples_per_second": 25.22,
|
7 |
+
"eval_steps_per_second": 0.79,
|
8 |
+
"eval_wer": 0.09092025712063107
|
9 |
+
}
|
optimizer.pt
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:c042e0a11d5efdeed407cea2d5b028d381a717b7d6b32c971c5452023a069a4c
|
3 |
-
size 2490346769
|
|
|
|
|
|
|
|
scheduler.pt → runs/Jun02_09-09-07_Niles/1654166101.6285756/events.out.tfevents.1654166101.Niles.467343.1
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:203c906e2b552322d9f6f9746255c2e220c579ca7a9d9ec74de2b313107b39e3
|
3 |
+
size 5020
|
rng_state.pth → runs/Jun02_09-09-07_Niles/events.out.tfevents.1654166101.Niles.467343.0
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:81b4c602d8b5a5e53abb965cc6fee7f8c60be2843c1320ebaf539ceb4a167986
|
3 |
+
size 968445
|
scaler.pt → runs/Jun02_09-09-07_Niles/events.out.tfevents.1655605972.Niles.467343.2
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6f74933ae1fc7058a67f68b8bbf8b5c78cc5085d417437262d7f01876c4d2d10
|
3 |
+
size 364
|
train_results.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 50.0,
|
3 |
+
"train_loss": 0.1478050081201038,
|
4 |
+
"train_runtime": 1439234.2576,
|
5 |
+
"train_samples": 455356,
|
6 |
+
"train_samples_per_second": 15.819,
|
7 |
+
"train_steps_per_second": 0.124
|
8 |
+
}
|
trainer_state.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.09092025712063107,
|
3 |
"best_model_checkpoint": "./wav2vec2-large-xlsr-53-german-cv9/checkpoint-160065",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -32421,11 +32421,3623 @@
|
|
32421 |
"eval_steps_per_second": 0.784,
|
32422 |
"eval_wer": 0.09092025712063107,
|
32423 |
"step": 160065
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
32424 |
}
|
32425 |
],
|
32426 |
"max_steps": 177850,
|
32427 |
"num_train_epochs": 50,
|
32428 |
-
"total_flos": 3.
|
32429 |
"trial_name": null,
|
32430 |
"trial_params": null
|
32431 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.09092025712063107,
|
3 |
"best_model_checkpoint": "./wav2vec2-large-xlsr-53-german-cv9/checkpoint-160065",
|
4 |
+
"epoch": 49.999859451862264,
|
5 |
+
"global_step": 177850,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
32421 |
"eval_steps_per_second": 0.784,
|
32422 |
"eval_wer": 0.09092025712063107,
|
32423 |
"step": 160065
|
32424 |
+
},
|
32425 |
+
{
|
32426 |
+
"epoch": 45.0,
|
32427 |
+
"learning_rate": 1.1140474182363415e-05,
|
32428 |
+
"loss": 0.0449,
|
32429 |
+
"step": 160080
|
32430 |
+
},
|
32431 |
+
{
|
32432 |
+
"epoch": 45.01,
|
32433 |
+
"learning_rate": 1.112173179645769e-05,
|
32434 |
+
"loss": 0.0429,
|
32435 |
+
"step": 160110
|
32436 |
+
},
|
32437 |
+
{
|
32438 |
+
"epoch": 45.02,
|
32439 |
+
"learning_rate": 1.1102989410551964e-05,
|
32440 |
+
"loss": 0.0432,
|
32441 |
+
"step": 160140
|
32442 |
+
},
|
32443 |
+
{
|
32444 |
+
"epoch": 45.03,
|
32445 |
+
"learning_rate": 1.1084247024646238e-05,
|
32446 |
+
"loss": 0.046,
|
32447 |
+
"step": 160170
|
32448 |
+
},
|
32449 |
+
{
|
32450 |
+
"epoch": 45.04,
|
32451 |
+
"learning_rate": 1.1065504638740512e-05,
|
32452 |
+
"loss": 0.0404,
|
32453 |
+
"step": 160200
|
32454 |
+
},
|
32455 |
+
{
|
32456 |
+
"epoch": 45.05,
|
32457 |
+
"learning_rate": 1.1046762252834786e-05,
|
32458 |
+
"loss": 0.0478,
|
32459 |
+
"step": 160230
|
32460 |
+
},
|
32461 |
+
{
|
32462 |
+
"epoch": 45.05,
|
32463 |
+
"learning_rate": 1.102801986692906e-05,
|
32464 |
+
"loss": 0.0436,
|
32465 |
+
"step": 160260
|
32466 |
+
},
|
32467 |
+
{
|
32468 |
+
"epoch": 45.06,
|
32469 |
+
"learning_rate": 1.1009277481023335e-05,
|
32470 |
+
"loss": 0.042,
|
32471 |
+
"step": 160290
|
32472 |
+
},
|
32473 |
+
{
|
32474 |
+
"epoch": 45.07,
|
32475 |
+
"learning_rate": 1.0990535095117609e-05,
|
32476 |
+
"loss": 0.0467,
|
32477 |
+
"step": 160320
|
32478 |
+
},
|
32479 |
+
{
|
32480 |
+
"epoch": 45.08,
|
32481 |
+
"learning_rate": 1.0971792709211883e-05,
|
32482 |
+
"loss": 0.0386,
|
32483 |
+
"step": 160350
|
32484 |
+
},
|
32485 |
+
{
|
32486 |
+
"epoch": 45.09,
|
32487 |
+
"learning_rate": 1.0953050323306157e-05,
|
32488 |
+
"loss": 0.0458,
|
32489 |
+
"step": 160380
|
32490 |
+
},
|
32491 |
+
{
|
32492 |
+
"epoch": 45.1,
|
32493 |
+
"learning_rate": 1.0934307937400431e-05,
|
32494 |
+
"loss": 0.0434,
|
32495 |
+
"step": 160410
|
32496 |
+
},
|
32497 |
+
{
|
32498 |
+
"epoch": 45.11,
|
32499 |
+
"learning_rate": 1.0915565551494706e-05,
|
32500 |
+
"loss": 0.0466,
|
32501 |
+
"step": 160440
|
32502 |
+
},
|
32503 |
+
{
|
32504 |
+
"epoch": 45.11,
|
32505 |
+
"learning_rate": 1.089682316558898e-05,
|
32506 |
+
"loss": 0.0487,
|
32507 |
+
"step": 160470
|
32508 |
+
},
|
32509 |
+
{
|
32510 |
+
"epoch": 45.12,
|
32511 |
+
"learning_rate": 1.0878080779683254e-05,
|
32512 |
+
"loss": 0.0414,
|
32513 |
+
"step": 160500
|
32514 |
+
},
|
32515 |
+
{
|
32516 |
+
"epoch": 45.13,
|
32517 |
+
"learning_rate": 1.0859338393777528e-05,
|
32518 |
+
"loss": 0.0465,
|
32519 |
+
"step": 160530
|
32520 |
+
},
|
32521 |
+
{
|
32522 |
+
"epoch": 45.14,
|
32523 |
+
"learning_rate": 1.0840596007871802e-05,
|
32524 |
+
"loss": 0.0424,
|
32525 |
+
"step": 160560
|
32526 |
+
},
|
32527 |
+
{
|
32528 |
+
"epoch": 45.15,
|
32529 |
+
"learning_rate": 1.0821853621966077e-05,
|
32530 |
+
"loss": 0.0402,
|
32531 |
+
"step": 160590
|
32532 |
+
},
|
32533 |
+
{
|
32534 |
+
"epoch": 45.16,
|
32535 |
+
"learning_rate": 1.080311123606035e-05,
|
32536 |
+
"loss": 0.0508,
|
32537 |
+
"step": 160620
|
32538 |
+
},
|
32539 |
+
{
|
32540 |
+
"epoch": 45.16,
|
32541 |
+
"learning_rate": 1.0784368850154625e-05,
|
32542 |
+
"loss": 0.0407,
|
32543 |
+
"step": 160650
|
32544 |
+
},
|
32545 |
+
{
|
32546 |
+
"epoch": 45.17,
|
32547 |
+
"learning_rate": 1.0765626464248899e-05,
|
32548 |
+
"loss": 0.0431,
|
32549 |
+
"step": 160680
|
32550 |
+
},
|
32551 |
+
{
|
32552 |
+
"epoch": 45.18,
|
32553 |
+
"learning_rate": 1.0746884078343173e-05,
|
32554 |
+
"loss": 0.0464,
|
32555 |
+
"step": 160710
|
32556 |
+
},
|
32557 |
+
{
|
32558 |
+
"epoch": 45.19,
|
32559 |
+
"learning_rate": 1.0728141692437447e-05,
|
32560 |
+
"loss": 0.0424,
|
32561 |
+
"step": 160740
|
32562 |
+
},
|
32563 |
+
{
|
32564 |
+
"epoch": 45.2,
|
32565 |
+
"learning_rate": 1.0709399306531722e-05,
|
32566 |
+
"loss": 0.0474,
|
32567 |
+
"step": 160770
|
32568 |
+
},
|
32569 |
+
{
|
32570 |
+
"epoch": 45.21,
|
32571 |
+
"learning_rate": 1.0690656920625996e-05,
|
32572 |
+
"loss": 0.0391,
|
32573 |
+
"step": 160800
|
32574 |
+
},
|
32575 |
+
{
|
32576 |
+
"epoch": 45.22,
|
32577 |
+
"learning_rate": 1.067191453472027e-05,
|
32578 |
+
"loss": 0.045,
|
32579 |
+
"step": 160830
|
32580 |
+
},
|
32581 |
+
{
|
32582 |
+
"epoch": 45.22,
|
32583 |
+
"learning_rate": 1.0653172148814544e-05,
|
32584 |
+
"loss": 0.0431,
|
32585 |
+
"step": 160860
|
32586 |
+
},
|
32587 |
+
{
|
32588 |
+
"epoch": 45.23,
|
32589 |
+
"learning_rate": 1.063442976290882e-05,
|
32590 |
+
"loss": 0.0418,
|
32591 |
+
"step": 160890
|
32592 |
+
},
|
32593 |
+
{
|
32594 |
+
"epoch": 45.24,
|
32595 |
+
"learning_rate": 1.0615687377003093e-05,
|
32596 |
+
"loss": 0.0493,
|
32597 |
+
"step": 160920
|
32598 |
+
},
|
32599 |
+
{
|
32600 |
+
"epoch": 45.25,
|
32601 |
+
"learning_rate": 1.0596944991097367e-05,
|
32602 |
+
"loss": 0.0404,
|
32603 |
+
"step": 160950
|
32604 |
+
},
|
32605 |
+
{
|
32606 |
+
"epoch": 45.26,
|
32607 |
+
"learning_rate": 1.0578202605191641e-05,
|
32608 |
+
"loss": 0.0482,
|
32609 |
+
"step": 160980
|
32610 |
+
},
|
32611 |
+
{
|
32612 |
+
"epoch": 45.27,
|
32613 |
+
"learning_rate": 1.0559460219285915e-05,
|
32614 |
+
"loss": 0.0464,
|
32615 |
+
"step": 161010
|
32616 |
+
},
|
32617 |
+
{
|
32618 |
+
"epoch": 45.27,
|
32619 |
+
"learning_rate": 1.054071783338019e-05,
|
32620 |
+
"loss": 0.0415,
|
32621 |
+
"step": 161040
|
32622 |
+
},
|
32623 |
+
{
|
32624 |
+
"epoch": 45.28,
|
32625 |
+
"learning_rate": 1.0521975447474464e-05,
|
32626 |
+
"loss": 0.0475,
|
32627 |
+
"step": 161070
|
32628 |
+
},
|
32629 |
+
{
|
32630 |
+
"epoch": 45.29,
|
32631 |
+
"learning_rate": 1.050323306156874e-05,
|
32632 |
+
"loss": 0.0405,
|
32633 |
+
"step": 161100
|
32634 |
+
},
|
32635 |
+
{
|
32636 |
+
"epoch": 45.3,
|
32637 |
+
"learning_rate": 1.0484490675663012e-05,
|
32638 |
+
"loss": 0.0459,
|
32639 |
+
"step": 161130
|
32640 |
+
},
|
32641 |
+
{
|
32642 |
+
"epoch": 45.31,
|
32643 |
+
"learning_rate": 1.0465748289757286e-05,
|
32644 |
+
"loss": 0.0424,
|
32645 |
+
"step": 161160
|
32646 |
+
},
|
32647 |
+
{
|
32648 |
+
"epoch": 45.32,
|
32649 |
+
"learning_rate": 1.044700590385156e-05,
|
32650 |
+
"loss": 0.041,
|
32651 |
+
"step": 161190
|
32652 |
+
},
|
32653 |
+
{
|
32654 |
+
"epoch": 45.32,
|
32655 |
+
"learning_rate": 1.0428263517945836e-05,
|
32656 |
+
"loss": 0.0477,
|
32657 |
+
"step": 161220
|
32658 |
+
},
|
32659 |
+
{
|
32660 |
+
"epoch": 45.33,
|
32661 |
+
"learning_rate": 1.0409521132040109e-05,
|
32662 |
+
"loss": 0.0398,
|
32663 |
+
"step": 161250
|
32664 |
+
},
|
32665 |
+
{
|
32666 |
+
"epoch": 45.34,
|
32667 |
+
"learning_rate": 1.0390778746134383e-05,
|
32668 |
+
"loss": 0.045,
|
32669 |
+
"step": 161280
|
32670 |
+
},
|
32671 |
+
{
|
32672 |
+
"epoch": 45.35,
|
32673 |
+
"learning_rate": 1.0372036360228659e-05,
|
32674 |
+
"loss": 0.0441,
|
32675 |
+
"step": 161310
|
32676 |
+
},
|
32677 |
+
{
|
32678 |
+
"epoch": 45.36,
|
32679 |
+
"learning_rate": 1.0353293974322931e-05,
|
32680 |
+
"loss": 0.0416,
|
32681 |
+
"step": 161340
|
32682 |
+
},
|
32683 |
+
{
|
32684 |
+
"epoch": 45.37,
|
32685 |
+
"learning_rate": 1.0334551588417205e-05,
|
32686 |
+
"loss": 0.0489,
|
32687 |
+
"step": 161370
|
32688 |
+
},
|
32689 |
+
{
|
32690 |
+
"epoch": 45.38,
|
32691 |
+
"learning_rate": 1.031580920251148e-05,
|
32692 |
+
"loss": 0.0406,
|
32693 |
+
"step": 161400
|
32694 |
+
},
|
32695 |
+
{
|
32696 |
+
"epoch": 45.38,
|
32697 |
+
"learning_rate": 1.0297066816605756e-05,
|
32698 |
+
"loss": 0.0474,
|
32699 |
+
"step": 161430
|
32700 |
+
},
|
32701 |
+
{
|
32702 |
+
"epoch": 45.39,
|
32703 |
+
"learning_rate": 1.0278324430700028e-05,
|
32704 |
+
"loss": 0.0453,
|
32705 |
+
"step": 161460
|
32706 |
+
},
|
32707 |
+
{
|
32708 |
+
"epoch": 45.4,
|
32709 |
+
"learning_rate": 1.0259582044794302e-05,
|
32710 |
+
"loss": 0.0429,
|
32711 |
+
"step": 161490
|
32712 |
+
},
|
32713 |
+
{
|
32714 |
+
"epoch": 45.41,
|
32715 |
+
"learning_rate": 1.0240839658888576e-05,
|
32716 |
+
"loss": 0.047,
|
32717 |
+
"step": 161520
|
32718 |
+
},
|
32719 |
+
{
|
32720 |
+
"epoch": 45.42,
|
32721 |
+
"learning_rate": 1.0222097272982852e-05,
|
32722 |
+
"loss": 0.0382,
|
32723 |
+
"step": 161550
|
32724 |
+
},
|
32725 |
+
{
|
32726 |
+
"epoch": 45.43,
|
32727 |
+
"learning_rate": 1.0203354887077125e-05,
|
32728 |
+
"loss": 0.0474,
|
32729 |
+
"step": 161580
|
32730 |
+
},
|
32731 |
+
{
|
32732 |
+
"epoch": 45.43,
|
32733 |
+
"learning_rate": 1.0184612501171399e-05,
|
32734 |
+
"loss": 0.044,
|
32735 |
+
"step": 161610
|
32736 |
+
},
|
32737 |
+
{
|
32738 |
+
"epoch": 45.44,
|
32739 |
+
"learning_rate": 1.0165870115265675e-05,
|
32740 |
+
"loss": 0.0432,
|
32741 |
+
"step": 161640
|
32742 |
+
},
|
32743 |
+
{
|
32744 |
+
"epoch": 45.45,
|
32745 |
+
"learning_rate": 1.0147127729359947e-05,
|
32746 |
+
"loss": 0.0456,
|
32747 |
+
"step": 161670
|
32748 |
+
},
|
32749 |
+
{
|
32750 |
+
"epoch": 45.46,
|
32751 |
+
"learning_rate": 1.0128385343454222e-05,
|
32752 |
+
"loss": 0.0401,
|
32753 |
+
"step": 161700
|
32754 |
+
},
|
32755 |
+
{
|
32756 |
+
"epoch": 45.47,
|
32757 |
+
"learning_rate": 1.0109642957548496e-05,
|
32758 |
+
"loss": 0.0461,
|
32759 |
+
"step": 161730
|
32760 |
+
},
|
32761 |
+
{
|
32762 |
+
"epoch": 45.48,
|
32763 |
+
"learning_rate": 1.0090900571642772e-05,
|
32764 |
+
"loss": 0.0423,
|
32765 |
+
"step": 161760
|
32766 |
+
},
|
32767 |
+
{
|
32768 |
+
"epoch": 45.48,
|
32769 |
+
"learning_rate": 1.0072158185737044e-05,
|
32770 |
+
"loss": 0.0451,
|
32771 |
+
"step": 161790
|
32772 |
+
},
|
32773 |
+
{
|
32774 |
+
"epoch": 45.49,
|
32775 |
+
"learning_rate": 1.0053415799831318e-05,
|
32776 |
+
"loss": 0.0462,
|
32777 |
+
"step": 161820
|
32778 |
+
},
|
32779 |
+
{
|
32780 |
+
"epoch": 45.5,
|
32781 |
+
"learning_rate": 1.0034673413925594e-05,
|
32782 |
+
"loss": 0.039,
|
32783 |
+
"step": 161850
|
32784 |
+
},
|
32785 |
+
{
|
32786 |
+
"epoch": 45.51,
|
32787 |
+
"learning_rate": 1.0015931028019868e-05,
|
32788 |
+
"loss": 0.0444,
|
32789 |
+
"step": 161880
|
32790 |
+
},
|
32791 |
+
{
|
32792 |
+
"epoch": 45.52,
|
32793 |
+
"learning_rate": 9.997188642114141e-06,
|
32794 |
+
"loss": 0.0398,
|
32795 |
+
"step": 161910
|
32796 |
+
},
|
32797 |
+
{
|
32798 |
+
"epoch": 45.53,
|
32799 |
+
"learning_rate": 9.978446256208415e-06,
|
32800 |
+
"loss": 0.0429,
|
32801 |
+
"step": 161940
|
32802 |
+
},
|
32803 |
+
{
|
32804 |
+
"epoch": 45.54,
|
32805 |
+
"learning_rate": 9.959703870302691e-06,
|
32806 |
+
"loss": 0.0514,
|
32807 |
+
"step": 161970
|
32808 |
+
},
|
32809 |
+
{
|
32810 |
+
"epoch": 45.54,
|
32811 |
+
"learning_rate": 9.940961484396964e-06,
|
32812 |
+
"loss": 0.0396,
|
32813 |
+
"step": 162000
|
32814 |
+
},
|
32815 |
+
{
|
32816 |
+
"epoch": 45.55,
|
32817 |
+
"learning_rate": 9.922219098491238e-06,
|
32818 |
+
"loss": 0.0469,
|
32819 |
+
"step": 162030
|
32820 |
+
},
|
32821 |
+
{
|
32822 |
+
"epoch": 45.56,
|
32823 |
+
"learning_rate": 9.903476712585514e-06,
|
32824 |
+
"loss": 0.041,
|
32825 |
+
"step": 162060
|
32826 |
+
},
|
32827 |
+
{
|
32828 |
+
"epoch": 45.57,
|
32829 |
+
"learning_rate": 9.884734326679788e-06,
|
32830 |
+
"loss": 0.0423,
|
32831 |
+
"step": 162090
|
32832 |
+
},
|
32833 |
+
{
|
32834 |
+
"epoch": 45.58,
|
32835 |
+
"learning_rate": 9.86599194077406e-06,
|
32836 |
+
"loss": 0.0495,
|
32837 |
+
"step": 162120
|
32838 |
+
},
|
32839 |
+
{
|
32840 |
+
"epoch": 45.59,
|
32841 |
+
"learning_rate": 9.847249554868334e-06,
|
32842 |
+
"loss": 0.0397,
|
32843 |
+
"step": 162150
|
32844 |
+
},
|
32845 |
+
{
|
32846 |
+
"epoch": 45.59,
|
32847 |
+
"learning_rate": 9.82850716896261e-06,
|
32848 |
+
"loss": 0.0457,
|
32849 |
+
"step": 162180
|
32850 |
+
},
|
32851 |
+
{
|
32852 |
+
"epoch": 45.6,
|
32853 |
+
"learning_rate": 9.809764783056883e-06,
|
32854 |
+
"loss": 0.0436,
|
32855 |
+
"step": 162210
|
32856 |
+
},
|
32857 |
+
{
|
32858 |
+
"epoch": 45.61,
|
32859 |
+
"learning_rate": 9.791022397151157e-06,
|
32860 |
+
"loss": 0.0412,
|
32861 |
+
"step": 162240
|
32862 |
+
},
|
32863 |
+
{
|
32864 |
+
"epoch": 45.62,
|
32865 |
+
"learning_rate": 9.772280011245433e-06,
|
32866 |
+
"loss": 0.0484,
|
32867 |
+
"step": 162270
|
32868 |
+
},
|
32869 |
+
{
|
32870 |
+
"epoch": 45.63,
|
32871 |
+
"learning_rate": 9.753537625339707e-06,
|
32872 |
+
"loss": 0.0388,
|
32873 |
+
"step": 162300
|
32874 |
+
},
|
32875 |
+
{
|
32876 |
+
"epoch": 45.64,
|
32877 |
+
"learning_rate": 9.73479523943398e-06,
|
32878 |
+
"loss": 0.0463,
|
32879 |
+
"step": 162330
|
32880 |
+
},
|
32881 |
+
{
|
32882 |
+
"epoch": 45.65,
|
32883 |
+
"learning_rate": 9.716052853528254e-06,
|
32884 |
+
"loss": 0.0428,
|
32885 |
+
"step": 162360
|
32886 |
+
},
|
32887 |
+
{
|
32888 |
+
"epoch": 45.65,
|
32889 |
+
"learning_rate": 9.69731046762253e-06,
|
32890 |
+
"loss": 0.0428,
|
32891 |
+
"step": 162390
|
32892 |
+
},
|
32893 |
+
{
|
32894 |
+
"epoch": 45.66,
|
32895 |
+
"learning_rate": 9.67919282791366e-06,
|
32896 |
+
"loss": 0.0507,
|
32897 |
+
"step": 162420
|
32898 |
+
},
|
32899 |
+
{
|
32900 |
+
"epoch": 45.67,
|
32901 |
+
"learning_rate": 9.660450442007935e-06,
|
32902 |
+
"loss": 0.041,
|
32903 |
+
"step": 162450
|
32904 |
+
},
|
32905 |
+
{
|
32906 |
+
"epoch": 45.68,
|
32907 |
+
"learning_rate": 9.641708056102209e-06,
|
32908 |
+
"loss": 0.0466,
|
32909 |
+
"step": 162480
|
32910 |
+
},
|
32911 |
+
{
|
32912 |
+
"epoch": 45.69,
|
32913 |
+
"learning_rate": 9.622965670196483e-06,
|
32914 |
+
"loss": 0.0427,
|
32915 |
+
"step": 162510
|
32916 |
+
},
|
32917 |
+
{
|
32918 |
+
"epoch": 45.7,
|
32919 |
+
"learning_rate": 9.604223284290757e-06,
|
32920 |
+
"loss": 0.0395,
|
32921 |
+
"step": 162540
|
32922 |
+
},
|
32923 |
+
{
|
32924 |
+
"epoch": 45.7,
|
32925 |
+
"learning_rate": 9.585480898385031e-06,
|
32926 |
+
"loss": 0.0468,
|
32927 |
+
"step": 162570
|
32928 |
+
},
|
32929 |
+
{
|
32930 |
+
"epoch": 45.71,
|
32931 |
+
"learning_rate": 9.566738512479306e-06,
|
32932 |
+
"loss": 0.0414,
|
32933 |
+
"step": 162600
|
32934 |
+
},
|
32935 |
+
{
|
32936 |
+
"epoch": 45.72,
|
32937 |
+
"learning_rate": 9.54799612657358e-06,
|
32938 |
+
"loss": 0.0458,
|
32939 |
+
"step": 162630
|
32940 |
+
},
|
32941 |
+
{
|
32942 |
+
"epoch": 45.73,
|
32943 |
+
"learning_rate": 9.529253740667854e-06,
|
32944 |
+
"loss": 0.0457,
|
32945 |
+
"step": 162660
|
32946 |
+
},
|
32947 |
+
{
|
32948 |
+
"epoch": 45.74,
|
32949 |
+
"learning_rate": 9.510511354762128e-06,
|
32950 |
+
"loss": 0.0424,
|
32951 |
+
"step": 162690
|
32952 |
+
},
|
32953 |
+
{
|
32954 |
+
"epoch": 45.75,
|
32955 |
+
"learning_rate": 9.491768968856402e-06,
|
32956 |
+
"loss": 0.0476,
|
32957 |
+
"step": 162720
|
32958 |
+
},
|
32959 |
+
{
|
32960 |
+
"epoch": 45.75,
|
32961 |
+
"learning_rate": 9.473026582950677e-06,
|
32962 |
+
"loss": 0.038,
|
32963 |
+
"step": 162750
|
32964 |
+
},
|
32965 |
+
{
|
32966 |
+
"epoch": 45.76,
|
32967 |
+
"learning_rate": 9.45428419704495e-06,
|
32968 |
+
"loss": 0.0455,
|
32969 |
+
"step": 162780
|
32970 |
+
},
|
32971 |
+
{
|
32972 |
+
"epoch": 45.77,
|
32973 |
+
"learning_rate": 9.435541811139225e-06,
|
32974 |
+
"loss": 0.0425,
|
32975 |
+
"step": 162810
|
32976 |
+
},
|
32977 |
+
{
|
32978 |
+
"epoch": 45.78,
|
32979 |
+
"learning_rate": 9.4167994252335e-06,
|
32980 |
+
"loss": 0.0477,
|
32981 |
+
"step": 162840
|
32982 |
+
},
|
32983 |
+
{
|
32984 |
+
"epoch": 45.79,
|
32985 |
+
"learning_rate": 9.398057039327773e-06,
|
32986 |
+
"loss": 0.0454,
|
32987 |
+
"step": 162870
|
32988 |
+
},
|
32989 |
+
{
|
32990 |
+
"epoch": 45.8,
|
32991 |
+
"learning_rate": 9.379314653422048e-06,
|
32992 |
+
"loss": 0.0376,
|
32993 |
+
"step": 162900
|
32994 |
+
},
|
32995 |
+
{
|
32996 |
+
"epoch": 45.81,
|
32997 |
+
"learning_rate": 9.360572267516322e-06,
|
32998 |
+
"loss": 0.0446,
|
32999 |
+
"step": 162930
|
33000 |
+
},
|
33001 |
+
{
|
33002 |
+
"epoch": 45.81,
|
33003 |
+
"learning_rate": 9.341829881610596e-06,
|
33004 |
+
"loss": 0.0435,
|
33005 |
+
"step": 162960
|
33006 |
+
},
|
33007 |
+
{
|
33008 |
+
"epoch": 45.82,
|
33009 |
+
"learning_rate": 9.32308749570487e-06,
|
33010 |
+
"loss": 0.0414,
|
33011 |
+
"step": 162990
|
33012 |
+
},
|
33013 |
+
{
|
33014 |
+
"epoch": 45.83,
|
33015 |
+
"learning_rate": 9.304345109799146e-06,
|
33016 |
+
"loss": 0.0479,
|
33017 |
+
"step": 163020
|
33018 |
+
},
|
33019 |
+
{
|
33020 |
+
"epoch": 45.84,
|
33021 |
+
"learning_rate": 9.285602723893418e-06,
|
33022 |
+
"loss": 0.0406,
|
33023 |
+
"step": 163050
|
33024 |
+
},
|
33025 |
+
{
|
33026 |
+
"epoch": 45.85,
|
33027 |
+
"learning_rate": 9.266860337987693e-06,
|
33028 |
+
"loss": 0.0436,
|
33029 |
+
"step": 163080
|
33030 |
+
},
|
33031 |
+
{
|
33032 |
+
"epoch": 45.86,
|
33033 |
+
"learning_rate": 9.248117952081967e-06,
|
33034 |
+
"loss": 0.0445,
|
33035 |
+
"step": 163110
|
33036 |
+
},
|
33037 |
+
{
|
33038 |
+
"epoch": 45.86,
|
33039 |
+
"learning_rate": 9.229375566176241e-06,
|
33040 |
+
"loss": 0.0416,
|
33041 |
+
"step": 163140
|
33042 |
+
},
|
33043 |
+
{
|
33044 |
+
"epoch": 45.87,
|
33045 |
+
"learning_rate": 9.210633180270515e-06,
|
33046 |
+
"loss": 0.0463,
|
33047 |
+
"step": 163170
|
33048 |
+
},
|
33049 |
+
{
|
33050 |
+
"epoch": 45.88,
|
33051 |
+
"learning_rate": 9.19189079436479e-06,
|
33052 |
+
"loss": 0.0397,
|
33053 |
+
"step": 163200
|
33054 |
+
},
|
33055 |
+
{
|
33056 |
+
"epoch": 45.89,
|
33057 |
+
"learning_rate": 9.173148408459064e-06,
|
33058 |
+
"loss": 0.0489,
|
33059 |
+
"step": 163230
|
33060 |
+
},
|
33061 |
+
{
|
33062 |
+
"epoch": 45.9,
|
33063 |
+
"learning_rate": 9.154406022553338e-06,
|
33064 |
+
"loss": 0.0442,
|
33065 |
+
"step": 163260
|
33066 |
+
},
|
33067 |
+
{
|
33068 |
+
"epoch": 45.91,
|
33069 |
+
"learning_rate": 9.135663636647612e-06,
|
33070 |
+
"loss": 0.0416,
|
33071 |
+
"step": 163290
|
33072 |
+
},
|
33073 |
+
{
|
33074 |
+
"epoch": 45.91,
|
33075 |
+
"learning_rate": 9.116921250741886e-06,
|
33076 |
+
"loss": 0.0478,
|
33077 |
+
"step": 163320
|
33078 |
+
},
|
33079 |
+
{
|
33080 |
+
"epoch": 45.92,
|
33081 |
+
"learning_rate": 9.098178864836162e-06,
|
33082 |
+
"loss": 0.0393,
|
33083 |
+
"step": 163350
|
33084 |
+
},
|
33085 |
+
{
|
33086 |
+
"epoch": 45.93,
|
33087 |
+
"learning_rate": 9.079436478930435e-06,
|
33088 |
+
"loss": 0.0456,
|
33089 |
+
"step": 163380
|
33090 |
+
},
|
33091 |
+
{
|
33092 |
+
"epoch": 45.94,
|
33093 |
+
"learning_rate": 9.060694093024709e-06,
|
33094 |
+
"loss": 0.045,
|
33095 |
+
"step": 163410
|
33096 |
+
},
|
33097 |
+
{
|
33098 |
+
"epoch": 45.95,
|
33099 |
+
"learning_rate": 9.041951707118983e-06,
|
33100 |
+
"loss": 0.0438,
|
33101 |
+
"step": 163440
|
33102 |
+
},
|
33103 |
+
{
|
33104 |
+
"epoch": 45.96,
|
33105 |
+
"learning_rate": 9.023209321213257e-06,
|
33106 |
+
"loss": 0.0497,
|
33107 |
+
"step": 163470
|
33108 |
+
},
|
33109 |
+
{
|
33110 |
+
"epoch": 45.97,
|
33111 |
+
"learning_rate": 9.004466935307531e-06,
|
33112 |
+
"loss": 0.0407,
|
33113 |
+
"step": 163500
|
33114 |
+
},
|
33115 |
+
{
|
33116 |
+
"epoch": 45.97,
|
33117 |
+
"learning_rate": 8.985724549401806e-06,
|
33118 |
+
"loss": 0.0465,
|
33119 |
+
"step": 163530
|
33120 |
+
},
|
33121 |
+
{
|
33122 |
+
"epoch": 45.98,
|
33123 |
+
"learning_rate": 8.966982163496081e-06,
|
33124 |
+
"loss": 0.0413,
|
33125 |
+
"step": 163560
|
33126 |
+
},
|
33127 |
+
{
|
33128 |
+
"epoch": 45.99,
|
33129 |
+
"learning_rate": 8.948239777590354e-06,
|
33130 |
+
"loss": 0.0459,
|
33131 |
+
"step": 163590
|
33132 |
+
},
|
33133 |
+
{
|
33134 |
+
"epoch": 46.0,
|
33135 |
+
"learning_rate": 8.929497391684628e-06,
|
33136 |
+
"loss": 0.0472,
|
33137 |
+
"step": 163620
|
33138 |
+
},
|
33139 |
+
{
|
33140 |
+
"epoch": 46.0,
|
33141 |
+
"eval_loss": 0.10915178805589676,
|
33142 |
+
"eval_runtime": 656.9625,
|
33143 |
+
"eval_samples_per_second": 24.405,
|
33144 |
+
"eval_steps_per_second": 0.764,
|
33145 |
+
"eval_wer": 0.09214171659410697,
|
33146 |
+
"step": 163622
|
33147 |
+
},
|
33148 |
+
{
|
33149 |
+
"epoch": 46.01,
|
33150 |
+
"learning_rate": 8.910755005778902e-06,
|
33151 |
+
"loss": 0.041,
|
33152 |
+
"step": 163650
|
33153 |
+
},
|
33154 |
+
{
|
33155 |
+
"epoch": 46.02,
|
33156 |
+
"learning_rate": 8.892012619873178e-06,
|
33157 |
+
"loss": 0.0449,
|
33158 |
+
"step": 163680
|
33159 |
+
},
|
33160 |
+
{
|
33161 |
+
"epoch": 46.02,
|
33162 |
+
"learning_rate": 8.87327023396745e-06,
|
33163 |
+
"loss": 0.0407,
|
33164 |
+
"step": 163710
|
33165 |
+
},
|
33166 |
+
{
|
33167 |
+
"epoch": 46.03,
|
33168 |
+
"learning_rate": 8.854527848061725e-06,
|
33169 |
+
"loss": 0.0458,
|
33170 |
+
"step": 163740
|
33171 |
+
},
|
33172 |
+
{
|
33173 |
+
"epoch": 46.04,
|
33174 |
+
"learning_rate": 8.835785462156e-06,
|
33175 |
+
"loss": 0.0444,
|
33176 |
+
"step": 163770
|
33177 |
+
},
|
33178 |
+
{
|
33179 |
+
"epoch": 46.05,
|
33180 |
+
"learning_rate": 8.817043076250273e-06,
|
33181 |
+
"loss": 0.0398,
|
33182 |
+
"step": 163800
|
33183 |
+
},
|
33184 |
+
{
|
33185 |
+
"epoch": 46.06,
|
33186 |
+
"learning_rate": 8.798300690344547e-06,
|
33187 |
+
"loss": 0.0483,
|
33188 |
+
"step": 163830
|
33189 |
+
},
|
33190 |
+
{
|
33191 |
+
"epoch": 46.07,
|
33192 |
+
"learning_rate": 8.779558304438822e-06,
|
33193 |
+
"loss": 0.0402,
|
33194 |
+
"step": 163860
|
33195 |
+
},
|
33196 |
+
{
|
33197 |
+
"epoch": 46.08,
|
33198 |
+
"learning_rate": 8.760815918533098e-06,
|
33199 |
+
"loss": 0.0444,
|
33200 |
+
"step": 163890
|
33201 |
+
},
|
33202 |
+
{
|
33203 |
+
"epoch": 46.08,
|
33204 |
+
"learning_rate": 8.74207353262737e-06,
|
33205 |
+
"loss": 0.0459,
|
33206 |
+
"step": 163920
|
33207 |
+
},
|
33208 |
+
{
|
33209 |
+
"epoch": 46.09,
|
33210 |
+
"learning_rate": 8.723331146721644e-06,
|
33211 |
+
"loss": 0.0388,
|
33212 |
+
"step": 163950
|
33213 |
+
},
|
33214 |
+
{
|
33215 |
+
"epoch": 46.1,
|
33216 |
+
"learning_rate": 8.70458876081592e-06,
|
33217 |
+
"loss": 0.048,
|
33218 |
+
"step": 163980
|
33219 |
+
},
|
33220 |
+
{
|
33221 |
+
"epoch": 46.11,
|
33222 |
+
"learning_rate": 8.685846374910193e-06,
|
33223 |
+
"loss": 0.0393,
|
33224 |
+
"step": 164010
|
33225 |
+
},
|
33226 |
+
{
|
33227 |
+
"epoch": 46.12,
|
33228 |
+
"learning_rate": 8.667103989004467e-06,
|
33229 |
+
"loss": 0.0411,
|
33230 |
+
"step": 164040
|
33231 |
+
},
|
33232 |
+
{
|
33233 |
+
"epoch": 46.13,
|
33234 |
+
"learning_rate": 8.648361603098741e-06,
|
33235 |
+
"loss": 0.0463,
|
33236 |
+
"step": 164070
|
33237 |
+
},
|
33238 |
+
{
|
33239 |
+
"epoch": 46.13,
|
33240 |
+
"learning_rate": 8.629619217193017e-06,
|
33241 |
+
"loss": 0.0403,
|
33242 |
+
"step": 164100
|
33243 |
+
},
|
33244 |
+
{
|
33245 |
+
"epoch": 46.14,
|
33246 |
+
"learning_rate": 8.61087683128729e-06,
|
33247 |
+
"loss": 0.0463,
|
33248 |
+
"step": 164130
|
33249 |
+
},
|
33250 |
+
{
|
33251 |
+
"epoch": 46.15,
|
33252 |
+
"learning_rate": 8.592134445381564e-06,
|
33253 |
+
"loss": 0.0421,
|
33254 |
+
"step": 164160
|
33255 |
+
},
|
33256 |
+
{
|
33257 |
+
"epoch": 46.16,
|
33258 |
+
"learning_rate": 8.573392059475838e-06,
|
33259 |
+
"loss": 0.0467,
|
33260 |
+
"step": 164190
|
33261 |
+
},
|
33262 |
+
{
|
33263 |
+
"epoch": 46.17,
|
33264 |
+
"learning_rate": 8.554649673570114e-06,
|
33265 |
+
"loss": 0.0442,
|
33266 |
+
"step": 164220
|
33267 |
+
},
|
33268 |
+
{
|
33269 |
+
"epoch": 46.18,
|
33270 |
+
"learning_rate": 8.535907287664386e-06,
|
33271 |
+
"loss": 0.0402,
|
33272 |
+
"step": 164250
|
33273 |
+
},
|
33274 |
+
{
|
33275 |
+
"epoch": 46.18,
|
33276 |
+
"learning_rate": 8.51716490175866e-06,
|
33277 |
+
"loss": 0.046,
|
33278 |
+
"step": 164280
|
33279 |
+
},
|
33280 |
+
{
|
33281 |
+
"epoch": 46.19,
|
33282 |
+
"learning_rate": 8.498422515852936e-06,
|
33283 |
+
"loss": 0.0392,
|
33284 |
+
"step": 164310
|
33285 |
+
},
|
33286 |
+
{
|
33287 |
+
"epoch": 46.2,
|
33288 |
+
"learning_rate": 8.479680129947209e-06,
|
33289 |
+
"loss": 0.0438,
|
33290 |
+
"step": 164340
|
33291 |
+
},
|
33292 |
+
{
|
33293 |
+
"epoch": 46.21,
|
33294 |
+
"learning_rate": 8.460937744041483e-06,
|
33295 |
+
"loss": 0.0473,
|
33296 |
+
"step": 164370
|
33297 |
+
},
|
33298 |
+
{
|
33299 |
+
"epoch": 46.22,
|
33300 |
+
"learning_rate": 8.442195358135757e-06,
|
33301 |
+
"loss": 0.0431,
|
33302 |
+
"step": 164400
|
33303 |
+
},
|
33304 |
+
{
|
33305 |
+
"epoch": 46.23,
|
33306 |
+
"learning_rate": 8.423452972230033e-06,
|
33307 |
+
"loss": 0.0452,
|
33308 |
+
"step": 164430
|
33309 |
+
},
|
33310 |
+
{
|
33311 |
+
"epoch": 46.24,
|
33312 |
+
"learning_rate": 8.404710586324305e-06,
|
33313 |
+
"loss": 0.0409,
|
33314 |
+
"step": 164460
|
33315 |
+
},
|
33316 |
+
{
|
33317 |
+
"epoch": 46.24,
|
33318 |
+
"learning_rate": 8.38596820041858e-06,
|
33319 |
+
"loss": 0.047,
|
33320 |
+
"step": 164490
|
33321 |
+
},
|
33322 |
+
{
|
33323 |
+
"epoch": 46.25,
|
33324 |
+
"learning_rate": 8.367225814512856e-06,
|
33325 |
+
"loss": 0.0439,
|
33326 |
+
"step": 164520
|
33327 |
+
},
|
33328 |
+
{
|
33329 |
+
"epoch": 46.26,
|
33330 |
+
"learning_rate": 8.34848342860713e-06,
|
33331 |
+
"loss": 0.0421,
|
33332 |
+
"step": 164550
|
33333 |
+
},
|
33334 |
+
{
|
33335 |
+
"epoch": 46.27,
|
33336 |
+
"learning_rate": 8.329741042701402e-06,
|
33337 |
+
"loss": 0.0489,
|
33338 |
+
"step": 164580
|
33339 |
+
},
|
33340 |
+
{
|
33341 |
+
"epoch": 46.28,
|
33342 |
+
"learning_rate": 8.310998656795676e-06,
|
33343 |
+
"loss": 0.04,
|
33344 |
+
"step": 164610
|
33345 |
+
},
|
33346 |
+
{
|
33347 |
+
"epoch": 46.29,
|
33348 |
+
"learning_rate": 8.292256270889952e-06,
|
33349 |
+
"loss": 0.0432,
|
33350 |
+
"step": 164640
|
33351 |
+
},
|
33352 |
+
{
|
33353 |
+
"epoch": 46.29,
|
33354 |
+
"learning_rate": 8.273513884984225e-06,
|
33355 |
+
"loss": 0.0481,
|
33356 |
+
"step": 164670
|
33357 |
+
},
|
33358 |
+
{
|
33359 |
+
"epoch": 46.3,
|
33360 |
+
"learning_rate": 8.254771499078499e-06,
|
33361 |
+
"loss": 0.0428,
|
33362 |
+
"step": 164700
|
33363 |
+
},
|
33364 |
+
{
|
33365 |
+
"epoch": 46.31,
|
33366 |
+
"learning_rate": 8.236029113172775e-06,
|
33367 |
+
"loss": 0.0482,
|
33368 |
+
"step": 164730
|
33369 |
+
},
|
33370 |
+
{
|
33371 |
+
"epoch": 46.32,
|
33372 |
+
"learning_rate": 8.217286727267049e-06,
|
33373 |
+
"loss": 0.042,
|
33374 |
+
"step": 164760
|
33375 |
+
},
|
33376 |
+
{
|
33377 |
+
"epoch": 46.33,
|
33378 |
+
"learning_rate": 8.198544341361322e-06,
|
33379 |
+
"loss": 0.0435,
|
33380 |
+
"step": 164790
|
33381 |
+
},
|
33382 |
+
{
|
33383 |
+
"epoch": 46.34,
|
33384 |
+
"learning_rate": 8.179801955455596e-06,
|
33385 |
+
"loss": 0.0469,
|
33386 |
+
"step": 164820
|
33387 |
+
},
|
33388 |
+
{
|
33389 |
+
"epoch": 46.35,
|
33390 |
+
"learning_rate": 8.161059569549872e-06,
|
33391 |
+
"loss": 0.0398,
|
33392 |
+
"step": 164850
|
33393 |
+
},
|
33394 |
+
{
|
33395 |
+
"epoch": 46.35,
|
33396 |
+
"learning_rate": 8.142317183644146e-06,
|
33397 |
+
"loss": 0.048,
|
33398 |
+
"step": 164880
|
33399 |
+
},
|
33400 |
+
{
|
33401 |
+
"epoch": 46.36,
|
33402 |
+
"learning_rate": 8.123574797738418e-06,
|
33403 |
+
"loss": 0.0394,
|
33404 |
+
"step": 164910
|
33405 |
+
},
|
33406 |
+
{
|
33407 |
+
"epoch": 46.37,
|
33408 |
+
"learning_rate": 8.10545715802955e-06,
|
33409 |
+
"loss": 0.0445,
|
33410 |
+
"step": 164940
|
33411 |
+
},
|
33412 |
+
{
|
33413 |
+
"epoch": 46.38,
|
33414 |
+
"learning_rate": 8.086714772123825e-06,
|
33415 |
+
"loss": 0.044,
|
33416 |
+
"step": 164970
|
33417 |
+
},
|
33418 |
+
{
|
33419 |
+
"epoch": 46.39,
|
33420 |
+
"learning_rate": 8.0679723862181e-06,
|
33421 |
+
"loss": 0.0404,
|
33422 |
+
"step": 165000
|
33423 |
+
},
|
33424 |
+
{
|
33425 |
+
"epoch": 46.4,
|
33426 |
+
"learning_rate": 8.049230000312373e-06,
|
33427 |
+
"loss": 0.0454,
|
33428 |
+
"step": 165030
|
33429 |
+
},
|
33430 |
+
{
|
33431 |
+
"epoch": 46.4,
|
33432 |
+
"learning_rate": 8.030487614406648e-06,
|
33433 |
+
"loss": 0.0392,
|
33434 |
+
"step": 165060
|
33435 |
+
},
|
33436 |
+
{
|
33437 |
+
"epoch": 46.41,
|
33438 |
+
"learning_rate": 8.011745228500922e-06,
|
33439 |
+
"loss": 0.0465,
|
33440 |
+
"step": 165090
|
33441 |
+
},
|
33442 |
+
{
|
33443 |
+
"epoch": 46.42,
|
33444 |
+
"learning_rate": 7.993002842595196e-06,
|
33445 |
+
"loss": 0.0429,
|
33446 |
+
"step": 165120
|
33447 |
+
},
|
33448 |
+
{
|
33449 |
+
"epoch": 46.43,
|
33450 |
+
"learning_rate": 7.97426045668947e-06,
|
33451 |
+
"loss": 0.0394,
|
33452 |
+
"step": 165150
|
33453 |
+
},
|
33454 |
+
{
|
33455 |
+
"epoch": 46.44,
|
33456 |
+
"learning_rate": 7.955518070783744e-06,
|
33457 |
+
"loss": 0.0443,
|
33458 |
+
"step": 165180
|
33459 |
+
},
|
33460 |
+
{
|
33461 |
+
"epoch": 46.45,
|
33462 |
+
"learning_rate": 7.936775684878019e-06,
|
33463 |
+
"loss": 0.0377,
|
33464 |
+
"step": 165210
|
33465 |
+
},
|
33466 |
+
{
|
33467 |
+
"epoch": 46.45,
|
33468 |
+
"learning_rate": 7.918033298972293e-06,
|
33469 |
+
"loss": 0.0425,
|
33470 |
+
"step": 165240
|
33471 |
+
},
|
33472 |
+
{
|
33473 |
+
"epoch": 46.46,
|
33474 |
+
"learning_rate": 7.899290913066567e-06,
|
33475 |
+
"loss": 0.0444,
|
33476 |
+
"step": 165270
|
33477 |
+
},
|
33478 |
+
{
|
33479 |
+
"epoch": 46.47,
|
33480 |
+
"learning_rate": 7.880548527160841e-06,
|
33481 |
+
"loss": 0.0401,
|
33482 |
+
"step": 165300
|
33483 |
+
},
|
33484 |
+
{
|
33485 |
+
"epoch": 46.48,
|
33486 |
+
"learning_rate": 7.861806141255115e-06,
|
33487 |
+
"loss": 0.0441,
|
33488 |
+
"step": 165330
|
33489 |
+
},
|
33490 |
+
{
|
33491 |
+
"epoch": 46.49,
|
33492 |
+
"learning_rate": 7.84306375534939e-06,
|
33493 |
+
"loss": 0.0437,
|
33494 |
+
"step": 165360
|
33495 |
+
},
|
33496 |
+
{
|
33497 |
+
"epoch": 46.5,
|
33498 |
+
"learning_rate": 7.824321369443664e-06,
|
33499 |
+
"loss": 0.0452,
|
33500 |
+
"step": 165390
|
33501 |
+
},
|
33502 |
+
{
|
33503 |
+
"epoch": 46.51,
|
33504 |
+
"learning_rate": 7.805578983537938e-06,
|
33505 |
+
"loss": 0.0437,
|
33506 |
+
"step": 165420
|
33507 |
+
},
|
33508 |
+
{
|
33509 |
+
"epoch": 46.51,
|
33510 |
+
"learning_rate": 7.786836597632212e-06,
|
33511 |
+
"loss": 0.042,
|
33512 |
+
"step": 165450
|
33513 |
+
},
|
33514 |
+
{
|
33515 |
+
"epoch": 46.52,
|
33516 |
+
"learning_rate": 7.768094211726486e-06,
|
33517 |
+
"loss": 0.0472,
|
33518 |
+
"step": 165480
|
33519 |
+
},
|
33520 |
+
{
|
33521 |
+
"epoch": 46.53,
|
33522 |
+
"learning_rate": 7.74935182582076e-06,
|
33523 |
+
"loss": 0.0394,
|
33524 |
+
"step": 165510
|
33525 |
+
},
|
33526 |
+
{
|
33527 |
+
"epoch": 46.54,
|
33528 |
+
"learning_rate": 7.730609439915035e-06,
|
33529 |
+
"loss": 0.0463,
|
33530 |
+
"step": 165540
|
33531 |
+
},
|
33532 |
+
{
|
33533 |
+
"epoch": 46.55,
|
33534 |
+
"learning_rate": 7.711867054009309e-06,
|
33535 |
+
"loss": 0.0452,
|
33536 |
+
"step": 165570
|
33537 |
+
},
|
33538 |
+
{
|
33539 |
+
"epoch": 46.56,
|
33540 |
+
"learning_rate": 7.693124668103583e-06,
|
33541 |
+
"loss": 0.0436,
|
33542 |
+
"step": 165600
|
33543 |
+
},
|
33544 |
+
{
|
33545 |
+
"epoch": 46.56,
|
33546 |
+
"learning_rate": 7.674382282197857e-06,
|
33547 |
+
"loss": 0.0459,
|
33548 |
+
"step": 165630
|
33549 |
+
},
|
33550 |
+
{
|
33551 |
+
"epoch": 46.57,
|
33552 |
+
"learning_rate": 7.655639896292131e-06,
|
33553 |
+
"loss": 0.0417,
|
33554 |
+
"step": 165660
|
33555 |
+
},
|
33556 |
+
{
|
33557 |
+
"epoch": 46.58,
|
33558 |
+
"learning_rate": 7.636897510386407e-06,
|
33559 |
+
"loss": 0.0445,
|
33560 |
+
"step": 165690
|
33561 |
+
},
|
33562 |
+
{
|
33563 |
+
"epoch": 46.59,
|
33564 |
+
"learning_rate": 7.618155124480681e-06,
|
33565 |
+
"loss": 0.0454,
|
33566 |
+
"step": 165720
|
33567 |
+
},
|
33568 |
+
{
|
33569 |
+
"epoch": 46.6,
|
33570 |
+
"learning_rate": 7.599412738574954e-06,
|
33571 |
+
"loss": 0.0425,
|
33572 |
+
"step": 165750
|
33573 |
+
},
|
33574 |
+
{
|
33575 |
+
"epoch": 46.61,
|
33576 |
+
"learning_rate": 7.580670352669228e-06,
|
33577 |
+
"loss": 0.0469,
|
33578 |
+
"step": 165780
|
33579 |
+
},
|
33580 |
+
{
|
33581 |
+
"epoch": 46.62,
|
33582 |
+
"learning_rate": 7.561927966763503e-06,
|
33583 |
+
"loss": 0.0398,
|
33584 |
+
"step": 165810
|
33585 |
+
},
|
33586 |
+
{
|
33587 |
+
"epoch": 46.62,
|
33588 |
+
"learning_rate": 7.543185580857777e-06,
|
33589 |
+
"loss": 0.0464,
|
33590 |
+
"step": 165840
|
33591 |
+
},
|
33592 |
+
{
|
33593 |
+
"epoch": 46.63,
|
33594 |
+
"learning_rate": 7.524443194952051e-06,
|
33595 |
+
"loss": 0.0442,
|
33596 |
+
"step": 165870
|
33597 |
+
},
|
33598 |
+
{
|
33599 |
+
"epoch": 46.64,
|
33600 |
+
"learning_rate": 7.505700809046325e-06,
|
33601 |
+
"loss": 0.04,
|
33602 |
+
"step": 165900
|
33603 |
+
},
|
33604 |
+
{
|
33605 |
+
"epoch": 46.65,
|
33606 |
+
"learning_rate": 7.4869584231406e-06,
|
33607 |
+
"loss": 0.0485,
|
33608 |
+
"step": 165930
|
33609 |
+
},
|
33610 |
+
{
|
33611 |
+
"epoch": 46.66,
|
33612 |
+
"learning_rate": 7.468216037234873e-06,
|
33613 |
+
"loss": 0.043,
|
33614 |
+
"step": 165960
|
33615 |
+
},
|
33616 |
+
{
|
33617 |
+
"epoch": 46.67,
|
33618 |
+
"learning_rate": 7.4494736513291475e-06,
|
33619 |
+
"loss": 0.0453,
|
33620 |
+
"step": 165990
|
33621 |
+
},
|
33622 |
+
{
|
33623 |
+
"epoch": 46.67,
|
33624 |
+
"learning_rate": 7.4307312654234225e-06,
|
33625 |
+
"loss": 0.0435,
|
33626 |
+
"step": 166020
|
33627 |
+
},
|
33628 |
+
{
|
33629 |
+
"epoch": 46.68,
|
33630 |
+
"learning_rate": 7.411988879517697e-06,
|
33631 |
+
"loss": 0.039,
|
33632 |
+
"step": 166050
|
33633 |
+
},
|
33634 |
+
{
|
33635 |
+
"epoch": 46.69,
|
33636 |
+
"learning_rate": 7.39324649361197e-06,
|
33637 |
+
"loss": 0.0441,
|
33638 |
+
"step": 166080
|
33639 |
+
},
|
33640 |
+
{
|
33641 |
+
"epoch": 46.7,
|
33642 |
+
"learning_rate": 7.374504107706244e-06,
|
33643 |
+
"loss": 0.0413,
|
33644 |
+
"step": 166110
|
33645 |
+
},
|
33646 |
+
{
|
33647 |
+
"epoch": 46.71,
|
33648 |
+
"learning_rate": 7.355761721800519e-06,
|
33649 |
+
"loss": 0.0438,
|
33650 |
+
"step": 166140
|
33651 |
+
},
|
33652 |
+
{
|
33653 |
+
"epoch": 46.72,
|
33654 |
+
"learning_rate": 7.3370193358947935e-06,
|
33655 |
+
"loss": 0.0486,
|
33656 |
+
"step": 166170
|
33657 |
+
},
|
33658 |
+
{
|
33659 |
+
"epoch": 46.72,
|
33660 |
+
"learning_rate": 7.318276949989067e-06,
|
33661 |
+
"loss": 0.0413,
|
33662 |
+
"step": 166200
|
33663 |
+
},
|
33664 |
+
{
|
33665 |
+
"epoch": 46.73,
|
33666 |
+
"learning_rate": 7.299534564083342e-06,
|
33667 |
+
"loss": 0.0467,
|
33668 |
+
"step": 166230
|
33669 |
+
},
|
33670 |
+
{
|
33671 |
+
"epoch": 46.74,
|
33672 |
+
"learning_rate": 7.280792178177616e-06,
|
33673 |
+
"loss": 0.0401,
|
33674 |
+
"step": 166260
|
33675 |
+
},
|
33676 |
+
{
|
33677 |
+
"epoch": 46.75,
|
33678 |
+
"learning_rate": 7.262049792271889e-06,
|
33679 |
+
"loss": 0.0453,
|
33680 |
+
"step": 166290
|
33681 |
+
},
|
33682 |
+
{
|
33683 |
+
"epoch": 46.76,
|
33684 |
+
"learning_rate": 7.243307406366164e-06,
|
33685 |
+
"loss": 0.0443,
|
33686 |
+
"step": 166320
|
33687 |
+
},
|
33688 |
+
{
|
33689 |
+
"epoch": 46.77,
|
33690 |
+
"learning_rate": 7.224565020460439e-06,
|
33691 |
+
"loss": 0.0385,
|
33692 |
+
"step": 166350
|
33693 |
+
},
|
33694 |
+
{
|
33695 |
+
"epoch": 46.78,
|
33696 |
+
"learning_rate": 7.205822634554713e-06,
|
33697 |
+
"loss": 0.046,
|
33698 |
+
"step": 166380
|
33699 |
+
},
|
33700 |
+
{
|
33701 |
+
"epoch": 46.78,
|
33702 |
+
"learning_rate": 7.187080248648986e-06,
|
33703 |
+
"loss": 0.0381,
|
33704 |
+
"step": 166410
|
33705 |
+
},
|
33706 |
+
{
|
33707 |
+
"epoch": 46.79,
|
33708 |
+
"learning_rate": 7.168337862743261e-06,
|
33709 |
+
"loss": 0.0435,
|
33710 |
+
"step": 166440
|
33711 |
+
},
|
33712 |
+
{
|
33713 |
+
"epoch": 46.8,
|
33714 |
+
"learning_rate": 7.149595476837535e-06,
|
33715 |
+
"loss": 0.0468,
|
33716 |
+
"step": 166470
|
33717 |
+
},
|
33718 |
+
{
|
33719 |
+
"epoch": 46.81,
|
33720 |
+
"learning_rate": 7.130853090931809e-06,
|
33721 |
+
"loss": 0.0435,
|
33722 |
+
"step": 166500
|
33723 |
+
},
|
33724 |
+
{
|
33725 |
+
"epoch": 46.82,
|
33726 |
+
"learning_rate": 7.112110705026083e-06,
|
33727 |
+
"loss": 0.0523,
|
33728 |
+
"step": 166530
|
33729 |
+
},
|
33730 |
+
{
|
33731 |
+
"epoch": 46.83,
|
33732 |
+
"learning_rate": 7.093368319120358e-06,
|
33733 |
+
"loss": 0.0386,
|
33734 |
+
"step": 166560
|
33735 |
+
},
|
33736 |
+
{
|
33737 |
+
"epoch": 46.83,
|
33738 |
+
"learning_rate": 7.074625933214632e-06,
|
33739 |
+
"loss": 0.0455,
|
33740 |
+
"step": 166590
|
33741 |
+
},
|
33742 |
+
{
|
33743 |
+
"epoch": 46.84,
|
33744 |
+
"learning_rate": 7.0558835473089055e-06,
|
33745 |
+
"loss": 0.0438,
|
33746 |
+
"step": 166620
|
33747 |
+
},
|
33748 |
+
{
|
33749 |
+
"epoch": 46.85,
|
33750 |
+
"learning_rate": 7.037141161403181e-06,
|
33751 |
+
"loss": 0.0387,
|
33752 |
+
"step": 166650
|
33753 |
+
},
|
33754 |
+
{
|
33755 |
+
"epoch": 46.86,
|
33756 |
+
"learning_rate": 7.018398775497455e-06,
|
33757 |
+
"loss": 0.0464,
|
33758 |
+
"step": 166680
|
33759 |
+
},
|
33760 |
+
{
|
33761 |
+
"epoch": 46.87,
|
33762 |
+
"learning_rate": 6.999656389591729e-06,
|
33763 |
+
"loss": 0.0417,
|
33764 |
+
"step": 166710
|
33765 |
+
},
|
33766 |
+
{
|
33767 |
+
"epoch": 46.88,
|
33768 |
+
"learning_rate": 6.980914003686002e-06,
|
33769 |
+
"loss": 0.0456,
|
33770 |
+
"step": 166740
|
33771 |
+
},
|
33772 |
+
{
|
33773 |
+
"epoch": 46.88,
|
33774 |
+
"learning_rate": 6.962171617780277e-06,
|
33775 |
+
"loss": 0.0444,
|
33776 |
+
"step": 166770
|
33777 |
+
},
|
33778 |
+
{
|
33779 |
+
"epoch": 46.89,
|
33780 |
+
"learning_rate": 6.9434292318745515e-06,
|
33781 |
+
"loss": 0.0398,
|
33782 |
+
"step": 166800
|
33783 |
+
},
|
33784 |
+
{
|
33785 |
+
"epoch": 46.9,
|
33786 |
+
"learning_rate": 6.924686845968825e-06,
|
33787 |
+
"loss": 0.0457,
|
33788 |
+
"step": 166830
|
33789 |
+
},
|
33790 |
+
{
|
33791 |
+
"epoch": 46.91,
|
33792 |
+
"learning_rate": 6.905944460063099e-06,
|
33793 |
+
"loss": 0.039,
|
33794 |
+
"step": 166860
|
33795 |
+
},
|
33796 |
+
{
|
33797 |
+
"epoch": 46.92,
|
33798 |
+
"learning_rate": 6.887202074157374e-06,
|
33799 |
+
"loss": 0.0428,
|
33800 |
+
"step": 166890
|
33801 |
+
},
|
33802 |
+
{
|
33803 |
+
"epoch": 46.93,
|
33804 |
+
"learning_rate": 6.868459688251648e-06,
|
33805 |
+
"loss": 0.0496,
|
33806 |
+
"step": 166920
|
33807 |
+
},
|
33808 |
+
{
|
33809 |
+
"epoch": 46.94,
|
33810 |
+
"learning_rate": 6.849717302345922e-06,
|
33811 |
+
"loss": 0.0405,
|
33812 |
+
"step": 166950
|
33813 |
+
},
|
33814 |
+
{
|
33815 |
+
"epoch": 46.94,
|
33816 |
+
"learning_rate": 6.830974916440197e-06,
|
33817 |
+
"loss": 0.0467,
|
33818 |
+
"step": 166980
|
33819 |
+
},
|
33820 |
+
{
|
33821 |
+
"epoch": 46.95,
|
33822 |
+
"learning_rate": 6.812232530534471e-06,
|
33823 |
+
"loss": 0.0402,
|
33824 |
+
"step": 167010
|
33825 |
+
},
|
33826 |
+
{
|
33827 |
+
"epoch": 46.96,
|
33828 |
+
"learning_rate": 6.7941148908256025e-06,
|
33829 |
+
"loss": 0.0422,
|
33830 |
+
"step": 167040
|
33831 |
+
},
|
33832 |
+
{
|
33833 |
+
"epoch": 46.97,
|
33834 |
+
"learning_rate": 6.775372504919876e-06,
|
33835 |
+
"loss": 0.0447,
|
33836 |
+
"step": 167070
|
33837 |
+
},
|
33838 |
+
{
|
33839 |
+
"epoch": 46.98,
|
33840 |
+
"learning_rate": 6.756630119014151e-06,
|
33841 |
+
"loss": 0.0419,
|
33842 |
+
"step": 167100
|
33843 |
+
},
|
33844 |
+
{
|
33845 |
+
"epoch": 46.99,
|
33846 |
+
"learning_rate": 6.737887733108425e-06,
|
33847 |
+
"loss": 0.0466,
|
33848 |
+
"step": 167130
|
33849 |
+
},
|
33850 |
+
{
|
33851 |
+
"epoch": 46.99,
|
33852 |
+
"learning_rate": 6.719145347202699e-06,
|
33853 |
+
"loss": 0.0414,
|
33854 |
+
"step": 167160
|
33855 |
+
},
|
33856 |
+
{
|
33857 |
+
"epoch": 47.0,
|
33858 |
+
"eval_loss": 0.11058283597230911,
|
33859 |
+
"eval_runtime": 643.2521,
|
33860 |
+
"eval_samples_per_second": 24.925,
|
33861 |
+
"eval_steps_per_second": 0.78,
|
33862 |
+
"eval_wer": 0.09220313075757783,
|
33863 |
+
"step": 167179
|
33864 |
+
},
|
33865 |
+
{
|
33866 |
+
"epoch": 47.0,
|
33867 |
+
"learning_rate": 6.700402961296974e-06,
|
33868 |
+
"loss": 0.0464,
|
33869 |
+
"step": 167190
|
33870 |
+
},
|
33871 |
+
{
|
33872 |
+
"epoch": 47.01,
|
33873 |
+
"learning_rate": 6.681660575391248e-06,
|
33874 |
+
"loss": 0.0417,
|
33875 |
+
"step": 167220
|
33876 |
+
},
|
33877 |
+
{
|
33878 |
+
"epoch": 47.02,
|
33879 |
+
"learning_rate": 6.662918189485522e-06,
|
33880 |
+
"loss": 0.0473,
|
33881 |
+
"step": 167250
|
33882 |
+
},
|
33883 |
+
{
|
33884 |
+
"epoch": 47.03,
|
33885 |
+
"learning_rate": 6.644175803579795e-06,
|
33886 |
+
"loss": 0.0478,
|
33887 |
+
"step": 167280
|
33888 |
+
},
|
33889 |
+
{
|
33890 |
+
"epoch": 47.04,
|
33891 |
+
"learning_rate": 6.625433417674071e-06,
|
33892 |
+
"loss": 0.0377,
|
33893 |
+
"step": 167310
|
33894 |
+
},
|
33895 |
+
{
|
33896 |
+
"epoch": 47.05,
|
33897 |
+
"learning_rate": 6.606691031768344e-06,
|
33898 |
+
"loss": 0.0459,
|
33899 |
+
"step": 167340
|
33900 |
+
},
|
33901 |
+
{
|
33902 |
+
"epoch": 47.05,
|
33903 |
+
"learning_rate": 6.587948645862619e-06,
|
33904 |
+
"loss": 0.0415,
|
33905 |
+
"step": 167370
|
33906 |
+
},
|
33907 |
+
{
|
33908 |
+
"epoch": 47.06,
|
33909 |
+
"learning_rate": 6.569206259956894e-06,
|
33910 |
+
"loss": 0.0398,
|
33911 |
+
"step": 167400
|
33912 |
+
},
|
33913 |
+
{
|
33914 |
+
"epoch": 47.07,
|
33915 |
+
"learning_rate": 6.550463874051167e-06,
|
33916 |
+
"loss": 0.0488,
|
33917 |
+
"step": 167430
|
33918 |
+
},
|
33919 |
+
{
|
33920 |
+
"epoch": 47.08,
|
33921 |
+
"learning_rate": 6.531721488145441e-06,
|
33922 |
+
"loss": 0.0384,
|
33923 |
+
"step": 167460
|
33924 |
+
},
|
33925 |
+
{
|
33926 |
+
"epoch": 47.09,
|
33927 |
+
"learning_rate": 6.5129791022397145e-06,
|
33928 |
+
"loss": 0.0447,
|
33929 |
+
"step": 167490
|
33930 |
+
},
|
33931 |
+
{
|
33932 |
+
"epoch": 47.1,
|
33933 |
+
"learning_rate": 6.49423671633399e-06,
|
33934 |
+
"loss": 0.0396,
|
33935 |
+
"step": 167520
|
33936 |
+
},
|
33937 |
+
{
|
33938 |
+
"epoch": 47.1,
|
33939 |
+
"learning_rate": 6.475494330428264e-06,
|
33940 |
+
"loss": 0.0448,
|
33941 |
+
"step": 167550
|
33942 |
+
},
|
33943 |
+
{
|
33944 |
+
"epoch": 47.11,
|
33945 |
+
"learning_rate": 6.456751944522538e-06,
|
33946 |
+
"loss": 0.0464,
|
33947 |
+
"step": 167580
|
33948 |
+
},
|
33949 |
+
{
|
33950 |
+
"epoch": 47.12,
|
33951 |
+
"learning_rate": 6.438009558616811e-06,
|
33952 |
+
"loss": 0.0386,
|
33953 |
+
"step": 167610
|
33954 |
+
},
|
33955 |
+
{
|
33956 |
+
"epoch": 47.13,
|
33957 |
+
"learning_rate": 6.419267172711087e-06,
|
33958 |
+
"loss": 0.045,
|
33959 |
+
"step": 167640
|
33960 |
+
},
|
33961 |
+
{
|
33962 |
+
"epoch": 47.14,
|
33963 |
+
"learning_rate": 6.4005247868053605e-06,
|
33964 |
+
"loss": 0.0409,
|
33965 |
+
"step": 167670
|
33966 |
+
},
|
33967 |
+
{
|
33968 |
+
"epoch": 47.15,
|
33969 |
+
"learning_rate": 6.381782400899635e-06,
|
33970 |
+
"loss": 0.0437,
|
33971 |
+
"step": 167700
|
33972 |
+
},
|
33973 |
+
{
|
33974 |
+
"epoch": 47.15,
|
33975 |
+
"learning_rate": 6.36304001499391e-06,
|
33976 |
+
"loss": 0.0507,
|
33977 |
+
"step": 167730
|
33978 |
+
},
|
33979 |
+
{
|
33980 |
+
"epoch": 47.16,
|
33981 |
+
"learning_rate": 6.344297629088183e-06,
|
33982 |
+
"loss": 0.0384,
|
33983 |
+
"step": 167760
|
33984 |
+
},
|
33985 |
+
{
|
33986 |
+
"epoch": 47.17,
|
33987 |
+
"learning_rate": 6.325555243182457e-06,
|
33988 |
+
"loss": 0.0464,
|
33989 |
+
"step": 167790
|
33990 |
+
},
|
33991 |
+
{
|
33992 |
+
"epoch": 47.18,
|
33993 |
+
"learning_rate": 6.306812857276731e-06,
|
33994 |
+
"loss": 0.0404,
|
33995 |
+
"step": 167820
|
33996 |
+
},
|
33997 |
+
{
|
33998 |
+
"epoch": 47.19,
|
33999 |
+
"learning_rate": 6.2880704713710065e-06,
|
34000 |
+
"loss": 0.0443,
|
34001 |
+
"step": 167850
|
34002 |
+
},
|
34003 |
+
{
|
34004 |
+
"epoch": 47.2,
|
34005 |
+
"learning_rate": 6.26932808546528e-06,
|
34006 |
+
"loss": 0.0477,
|
34007 |
+
"step": 167880
|
34008 |
+
},
|
34009 |
+
{
|
34010 |
+
"epoch": 47.21,
|
34011 |
+
"learning_rate": 6.250585699559554e-06,
|
34012 |
+
"loss": 0.0374,
|
34013 |
+
"step": 167910
|
34014 |
+
},
|
34015 |
+
{
|
34016 |
+
"epoch": 47.21,
|
34017 |
+
"learning_rate": 6.231843313653828e-06,
|
34018 |
+
"loss": 0.0461,
|
34019 |
+
"step": 167940
|
34020 |
+
},
|
34021 |
+
{
|
34022 |
+
"epoch": 47.22,
|
34023 |
+
"learning_rate": 6.213100927748102e-06,
|
34024 |
+
"loss": 0.0431,
|
34025 |
+
"step": 167970
|
34026 |
+
},
|
34027 |
+
{
|
34028 |
+
"epoch": 47.23,
|
34029 |
+
"learning_rate": 6.194358541842377e-06,
|
34030 |
+
"loss": 0.0479,
|
34031 |
+
"step": 168000
|
34032 |
+
},
|
34033 |
+
{
|
34034 |
+
"epoch": 47.24,
|
34035 |
+
"learning_rate": 6.175616155936651e-06,
|
34036 |
+
"loss": 0.0479,
|
34037 |
+
"step": 168030
|
34038 |
+
},
|
34039 |
+
{
|
34040 |
+
"epoch": 47.25,
|
34041 |
+
"learning_rate": 6.156873770030925e-06,
|
34042 |
+
"loss": 0.0379,
|
34043 |
+
"step": 168060
|
34044 |
+
},
|
34045 |
+
{
|
34046 |
+
"epoch": 47.26,
|
34047 |
+
"learning_rate": 6.138131384125199e-06,
|
34048 |
+
"loss": 0.044,
|
34049 |
+
"step": 168090
|
34050 |
+
},
|
34051 |
+
{
|
34052 |
+
"epoch": 47.26,
|
34053 |
+
"learning_rate": 6.119388998219474e-06,
|
34054 |
+
"loss": 0.0413,
|
34055 |
+
"step": 168120
|
34056 |
+
},
|
34057 |
+
{
|
34058 |
+
"epoch": 47.27,
|
34059 |
+
"learning_rate": 6.1006466123137475e-06,
|
34060 |
+
"loss": 0.0432,
|
34061 |
+
"step": 168150
|
34062 |
+
},
|
34063 |
+
{
|
34064 |
+
"epoch": 47.28,
|
34065 |
+
"learning_rate": 6.081904226408023e-06,
|
34066 |
+
"loss": 0.0465,
|
34067 |
+
"step": 168180
|
34068 |
+
},
|
34069 |
+
{
|
34070 |
+
"epoch": 47.29,
|
34071 |
+
"learning_rate": 6.063161840502296e-06,
|
34072 |
+
"loss": 0.0373,
|
34073 |
+
"step": 168210
|
34074 |
+
},
|
34075 |
+
{
|
34076 |
+
"epoch": 47.3,
|
34077 |
+
"learning_rate": 6.044419454596571e-06,
|
34078 |
+
"loss": 0.0482,
|
34079 |
+
"step": 168240
|
34080 |
+
},
|
34081 |
+
{
|
34082 |
+
"epoch": 47.31,
|
34083 |
+
"learning_rate": 6.025677068690844e-06,
|
34084 |
+
"loss": 0.0423,
|
34085 |
+
"step": 168270
|
34086 |
+
},
|
34087 |
+
{
|
34088 |
+
"epoch": 47.32,
|
34089 |
+
"learning_rate": 6.0069346827851185e-06,
|
34090 |
+
"loss": 0.0453,
|
34091 |
+
"step": 168300
|
34092 |
+
},
|
34093 |
+
{
|
34094 |
+
"epoch": 47.32,
|
34095 |
+
"learning_rate": 5.988192296879393e-06,
|
34096 |
+
"loss": 0.0461,
|
34097 |
+
"step": 168330
|
34098 |
+
},
|
34099 |
+
{
|
34100 |
+
"epoch": 47.33,
|
34101 |
+
"learning_rate": 5.969449910973667e-06,
|
34102 |
+
"loss": 0.0361,
|
34103 |
+
"step": 168360
|
34104 |
+
},
|
34105 |
+
{
|
34106 |
+
"epoch": 47.34,
|
34107 |
+
"learning_rate": 5.950707525067942e-06,
|
34108 |
+
"loss": 0.0475,
|
34109 |
+
"step": 168390
|
34110 |
+
},
|
34111 |
+
{
|
34112 |
+
"epoch": 47.35,
|
34113 |
+
"learning_rate": 5.931965139162215e-06,
|
34114 |
+
"loss": 0.0415,
|
34115 |
+
"step": 168420
|
34116 |
+
},
|
34117 |
+
{
|
34118 |
+
"epoch": 47.36,
|
34119 |
+
"learning_rate": 5.91322275325649e-06,
|
34120 |
+
"loss": 0.0456,
|
34121 |
+
"step": 168450
|
34122 |
+
},
|
34123 |
+
{
|
34124 |
+
"epoch": 47.37,
|
34125 |
+
"learning_rate": 5.894480367350764e-06,
|
34126 |
+
"loss": 0.0492,
|
34127 |
+
"step": 168480
|
34128 |
+
},
|
34129 |
+
{
|
34130 |
+
"epoch": 47.37,
|
34131 |
+
"learning_rate": 5.875737981445039e-06,
|
34132 |
+
"loss": 0.0353,
|
34133 |
+
"step": 168510
|
34134 |
+
},
|
34135 |
+
{
|
34136 |
+
"epoch": 47.38,
|
34137 |
+
"learning_rate": 5.856995595539312e-06,
|
34138 |
+
"loss": 0.0436,
|
34139 |
+
"step": 168540
|
34140 |
+
},
|
34141 |
+
{
|
34142 |
+
"epoch": 47.39,
|
34143 |
+
"learning_rate": 5.838253209633587e-06,
|
34144 |
+
"loss": 0.0414,
|
34145 |
+
"step": 168570
|
34146 |
+
},
|
34147 |
+
{
|
34148 |
+
"epoch": 47.4,
|
34149 |
+
"learning_rate": 5.819510823727861e-06,
|
34150 |
+
"loss": 0.0415,
|
34151 |
+
"step": 168600
|
34152 |
+
},
|
34153 |
+
{
|
34154 |
+
"epoch": 47.41,
|
34155 |
+
"learning_rate": 5.800768437822135e-06,
|
34156 |
+
"loss": 0.0481,
|
34157 |
+
"step": 168630
|
34158 |
+
},
|
34159 |
+
{
|
34160 |
+
"epoch": 47.42,
|
34161 |
+
"learning_rate": 5.78202605191641e-06,
|
34162 |
+
"loss": 0.0374,
|
34163 |
+
"step": 168660
|
34164 |
+
},
|
34165 |
+
{
|
34166 |
+
"epoch": 47.42,
|
34167 |
+
"learning_rate": 5.763283666010683e-06,
|
34168 |
+
"loss": 0.0474,
|
34169 |
+
"step": 168690
|
34170 |
+
},
|
34171 |
+
{
|
34172 |
+
"epoch": 47.43,
|
34173 |
+
"learning_rate": 5.744541280104958e-06,
|
34174 |
+
"loss": 0.0409,
|
34175 |
+
"step": 168720
|
34176 |
+
},
|
34177 |
+
{
|
34178 |
+
"epoch": 47.44,
|
34179 |
+
"learning_rate": 5.725798894199231e-06,
|
34180 |
+
"loss": 0.0399,
|
34181 |
+
"step": 168750
|
34182 |
+
},
|
34183 |
+
{
|
34184 |
+
"epoch": 47.45,
|
34185 |
+
"learning_rate": 5.707056508293506e-06,
|
34186 |
+
"loss": 0.046,
|
34187 |
+
"step": 168780
|
34188 |
+
},
|
34189 |
+
{
|
34190 |
+
"epoch": 47.46,
|
34191 |
+
"learning_rate": 5.68831412238778e-06,
|
34192 |
+
"loss": 0.0394,
|
34193 |
+
"step": 168810
|
34194 |
+
},
|
34195 |
+
{
|
34196 |
+
"epoch": 47.47,
|
34197 |
+
"learning_rate": 5.669571736482055e-06,
|
34198 |
+
"loss": 0.045,
|
34199 |
+
"step": 168840
|
34200 |
+
},
|
34201 |
+
{
|
34202 |
+
"epoch": 47.48,
|
34203 |
+
"learning_rate": 5.650829350576329e-06,
|
34204 |
+
"loss": 0.043,
|
34205 |
+
"step": 168870
|
34206 |
+
},
|
34207 |
+
{
|
34208 |
+
"epoch": 47.48,
|
34209 |
+
"learning_rate": 5.632086964670602e-06,
|
34210 |
+
"loss": 0.041,
|
34211 |
+
"step": 168900
|
34212 |
+
},
|
34213 |
+
{
|
34214 |
+
"epoch": 47.49,
|
34215 |
+
"learning_rate": 5.613344578764877e-06,
|
34216 |
+
"loss": 0.0455,
|
34217 |
+
"step": 168930
|
34218 |
+
},
|
34219 |
+
{
|
34220 |
+
"epoch": 47.5,
|
34221 |
+
"learning_rate": 5.594602192859151e-06,
|
34222 |
+
"loss": 0.0376,
|
34223 |
+
"step": 168960
|
34224 |
+
},
|
34225 |
+
{
|
34226 |
+
"epoch": 47.51,
|
34227 |
+
"learning_rate": 5.575859806953426e-06,
|
34228 |
+
"loss": 0.0485,
|
34229 |
+
"step": 168990
|
34230 |
+
},
|
34231 |
+
{
|
34232 |
+
"epoch": 47.52,
|
34233 |
+
"learning_rate": 5.557117421047699e-06,
|
34234 |
+
"loss": 0.0398,
|
34235 |
+
"step": 169020
|
34236 |
+
},
|
34237 |
+
{
|
34238 |
+
"epoch": 47.53,
|
34239 |
+
"learning_rate": 5.538375035141974e-06,
|
34240 |
+
"loss": 0.0455,
|
34241 |
+
"step": 169050
|
34242 |
+
},
|
34243 |
+
{
|
34244 |
+
"epoch": 47.53,
|
34245 |
+
"learning_rate": 5.519632649236248e-06,
|
34246 |
+
"loss": 0.0467,
|
34247 |
+
"step": 169080
|
34248 |
+
},
|
34249 |
+
{
|
34250 |
+
"epoch": 47.54,
|
34251 |
+
"learning_rate": 5.5008902633305225e-06,
|
34252 |
+
"loss": 0.0384,
|
34253 |
+
"step": 169110
|
34254 |
+
},
|
34255 |
+
{
|
34256 |
+
"epoch": 47.55,
|
34257 |
+
"learning_rate": 5.482147877424797e-06,
|
34258 |
+
"loss": 0.0453,
|
34259 |
+
"step": 169140
|
34260 |
+
},
|
34261 |
+
{
|
34262 |
+
"epoch": 47.56,
|
34263 |
+
"learning_rate": 5.463405491519071e-06,
|
34264 |
+
"loss": 0.039,
|
34265 |
+
"step": 169170
|
34266 |
+
},
|
34267 |
+
{
|
34268 |
+
"epoch": 47.57,
|
34269 |
+
"learning_rate": 5.444663105613345e-06,
|
34270 |
+
"loss": 0.0439,
|
34271 |
+
"step": 169200
|
34272 |
+
},
|
34273 |
+
{
|
34274 |
+
"epoch": 47.58,
|
34275 |
+
"learning_rate": 5.425920719707618e-06,
|
34276 |
+
"loss": 0.0464,
|
34277 |
+
"step": 169230
|
34278 |
+
},
|
34279 |
+
{
|
34280 |
+
"epoch": 47.58,
|
34281 |
+
"learning_rate": 5.4071783338018934e-06,
|
34282 |
+
"loss": 0.036,
|
34283 |
+
"step": 169260
|
34284 |
+
},
|
34285 |
+
{
|
34286 |
+
"epoch": 47.59,
|
34287 |
+
"learning_rate": 5.388435947896167e-06,
|
34288 |
+
"loss": 0.0471,
|
34289 |
+
"step": 169290
|
34290 |
+
},
|
34291 |
+
{
|
34292 |
+
"epoch": 47.6,
|
34293 |
+
"learning_rate": 5.369693561990442e-06,
|
34294 |
+
"loss": 0.0394,
|
34295 |
+
"step": 169320
|
34296 |
+
},
|
34297 |
+
{
|
34298 |
+
"epoch": 47.61,
|
34299 |
+
"learning_rate": 5.3515759222815735e-06,
|
34300 |
+
"loss": 0.0413,
|
34301 |
+
"step": 169350
|
34302 |
+
},
|
34303 |
+
{
|
34304 |
+
"epoch": 47.62,
|
34305 |
+
"learning_rate": 5.332833536375848e-06,
|
34306 |
+
"loss": 0.0453,
|
34307 |
+
"step": 169380
|
34308 |
+
},
|
34309 |
+
{
|
34310 |
+
"epoch": 47.63,
|
34311 |
+
"learning_rate": 5.314091150470122e-06,
|
34312 |
+
"loss": 0.0386,
|
34313 |
+
"step": 169410
|
34314 |
+
},
|
34315 |
+
{
|
34316 |
+
"epoch": 47.64,
|
34317 |
+
"learning_rate": 5.295348764564396e-06,
|
34318 |
+
"loss": 0.0478,
|
34319 |
+
"step": 169440
|
34320 |
+
},
|
34321 |
+
{
|
34322 |
+
"epoch": 47.64,
|
34323 |
+
"learning_rate": 5.27660637865867e-06,
|
34324 |
+
"loss": 0.0398,
|
34325 |
+
"step": 169470
|
34326 |
+
},
|
34327 |
+
{
|
34328 |
+
"epoch": 47.65,
|
34329 |
+
"learning_rate": 5.2578639927529444e-06,
|
34330 |
+
"loss": 0.0428,
|
34331 |
+
"step": 169500
|
34332 |
+
},
|
34333 |
+
{
|
34334 |
+
"epoch": 47.66,
|
34335 |
+
"learning_rate": 5.239121606847219e-06,
|
34336 |
+
"loss": 0.047,
|
34337 |
+
"step": 169530
|
34338 |
+
},
|
34339 |
+
{
|
34340 |
+
"epoch": 47.67,
|
34341 |
+
"learning_rate": 5.220379220941493e-06,
|
34342 |
+
"loss": 0.0369,
|
34343 |
+
"step": 169560
|
34344 |
+
},
|
34345 |
+
{
|
34346 |
+
"epoch": 47.68,
|
34347 |
+
"learning_rate": 5.201636835035767e-06,
|
34348 |
+
"loss": 0.0455,
|
34349 |
+
"step": 169590
|
34350 |
+
},
|
34351 |
+
{
|
34352 |
+
"epoch": 47.69,
|
34353 |
+
"learning_rate": 5.182894449130041e-06,
|
34354 |
+
"loss": 0.0414,
|
34355 |
+
"step": 169620
|
34356 |
+
},
|
34357 |
+
{
|
34358 |
+
"epoch": 47.69,
|
34359 |
+
"learning_rate": 5.164152063224315e-06,
|
34360 |
+
"loss": 0.0421,
|
34361 |
+
"step": 169650
|
34362 |
+
},
|
34363 |
+
{
|
34364 |
+
"epoch": 47.7,
|
34365 |
+
"learning_rate": 5.14540967731859e-06,
|
34366 |
+
"loss": 0.0445,
|
34367 |
+
"step": 169680
|
34368 |
+
},
|
34369 |
+
{
|
34370 |
+
"epoch": 47.71,
|
34371 |
+
"learning_rate": 5.126667291412864e-06,
|
34372 |
+
"loss": 0.0384,
|
34373 |
+
"step": 169710
|
34374 |
+
},
|
34375 |
+
{
|
34376 |
+
"epoch": 47.72,
|
34377 |
+
"learning_rate": 5.107924905507138e-06,
|
34378 |
+
"loss": 0.0456,
|
34379 |
+
"step": 169740
|
34380 |
+
},
|
34381 |
+
{
|
34382 |
+
"epoch": 47.73,
|
34383 |
+
"learning_rate": 5.089182519601412e-06,
|
34384 |
+
"loss": 0.0426,
|
34385 |
+
"step": 169770
|
34386 |
+
},
|
34387 |
+
{
|
34388 |
+
"epoch": 47.74,
|
34389 |
+
"learning_rate": 5.070440133695686e-06,
|
34390 |
+
"loss": 0.0431,
|
34391 |
+
"step": 169800
|
34392 |
+
},
|
34393 |
+
{
|
34394 |
+
"epoch": 47.75,
|
34395 |
+
"learning_rate": 5.0516977477899605e-06,
|
34396 |
+
"loss": 0.0463,
|
34397 |
+
"step": 169830
|
34398 |
+
},
|
34399 |
+
{
|
34400 |
+
"epoch": 47.75,
|
34401 |
+
"learning_rate": 5.032955361884235e-06,
|
34402 |
+
"loss": 0.036,
|
34403 |
+
"step": 169860
|
34404 |
+
},
|
34405 |
+
{
|
34406 |
+
"epoch": 47.76,
|
34407 |
+
"learning_rate": 5.014212975978509e-06,
|
34408 |
+
"loss": 0.0431,
|
34409 |
+
"step": 169890
|
34410 |
+
},
|
34411 |
+
{
|
34412 |
+
"epoch": 47.77,
|
34413 |
+
"learning_rate": 4.995470590072783e-06,
|
34414 |
+
"loss": 0.0406,
|
34415 |
+
"step": 169920
|
34416 |
+
},
|
34417 |
+
{
|
34418 |
+
"epoch": 47.78,
|
34419 |
+
"learning_rate": 4.976728204167057e-06,
|
34420 |
+
"loss": 0.0439,
|
34421 |
+
"step": 169950
|
34422 |
+
},
|
34423 |
+
{
|
34424 |
+
"epoch": 47.79,
|
34425 |
+
"learning_rate": 4.9579858182613315e-06,
|
34426 |
+
"loss": 0.0477,
|
34427 |
+
"step": 169980
|
34428 |
+
},
|
34429 |
+
{
|
34430 |
+
"epoch": 47.8,
|
34431 |
+
"learning_rate": 4.939243432355606e-06,
|
34432 |
+
"loss": 0.0387,
|
34433 |
+
"step": 170010
|
34434 |
+
},
|
34435 |
+
{
|
34436 |
+
"epoch": 47.8,
|
34437 |
+
"learning_rate": 4.92050104644988e-06,
|
34438 |
+
"loss": 0.0463,
|
34439 |
+
"step": 170040
|
34440 |
+
},
|
34441 |
+
{
|
34442 |
+
"epoch": 47.81,
|
34443 |
+
"learning_rate": 4.901758660544154e-06,
|
34444 |
+
"loss": 0.0409,
|
34445 |
+
"step": 170070
|
34446 |
+
},
|
34447 |
+
{
|
34448 |
+
"epoch": 47.82,
|
34449 |
+
"learning_rate": 4.883016274638428e-06,
|
34450 |
+
"loss": 0.0425,
|
34451 |
+
"step": 170100
|
34452 |
+
},
|
34453 |
+
{
|
34454 |
+
"epoch": 47.83,
|
34455 |
+
"learning_rate": 4.8642738887327024e-06,
|
34456 |
+
"loss": 0.0483,
|
34457 |
+
"step": 170130
|
34458 |
+
},
|
34459 |
+
{
|
34460 |
+
"epoch": 47.84,
|
34461 |
+
"learning_rate": 4.845531502826977e-06,
|
34462 |
+
"loss": 0.0387,
|
34463 |
+
"step": 170160
|
34464 |
+
},
|
34465 |
+
{
|
34466 |
+
"epoch": 47.85,
|
34467 |
+
"learning_rate": 4.826789116921251e-06,
|
34468 |
+
"loss": 0.0458,
|
34469 |
+
"step": 170190
|
34470 |
+
},
|
34471 |
+
{
|
34472 |
+
"epoch": 47.85,
|
34473 |
+
"learning_rate": 4.808046731015525e-06,
|
34474 |
+
"loss": 0.041,
|
34475 |
+
"step": 170220
|
34476 |
+
},
|
34477 |
+
{
|
34478 |
+
"epoch": 47.86,
|
34479 |
+
"learning_rate": 4.789304345109799e-06,
|
34480 |
+
"loss": 0.0432,
|
34481 |
+
"step": 170250
|
34482 |
+
},
|
34483 |
+
{
|
34484 |
+
"epoch": 47.87,
|
34485 |
+
"learning_rate": 4.770561959204073e-06,
|
34486 |
+
"loss": 0.046,
|
34487 |
+
"step": 170280
|
34488 |
+
},
|
34489 |
+
{
|
34490 |
+
"epoch": 47.88,
|
34491 |
+
"learning_rate": 4.7518195732983484e-06,
|
34492 |
+
"loss": 0.0368,
|
34493 |
+
"step": 170310
|
34494 |
+
},
|
34495 |
+
{
|
34496 |
+
"epoch": 47.89,
|
34497 |
+
"learning_rate": 4.733077187392622e-06,
|
34498 |
+
"loss": 0.0445,
|
34499 |
+
"step": 170340
|
34500 |
+
},
|
34501 |
+
{
|
34502 |
+
"epoch": 47.9,
|
34503 |
+
"learning_rate": 4.714334801486897e-06,
|
34504 |
+
"loss": 0.0409,
|
34505 |
+
"step": 170370
|
34506 |
+
},
|
34507 |
+
{
|
34508 |
+
"epoch": 47.91,
|
34509 |
+
"learning_rate": 4.69559241558117e-06,
|
34510 |
+
"loss": 0.0428,
|
34511 |
+
"step": 170400
|
34512 |
+
},
|
34513 |
+
{
|
34514 |
+
"epoch": 47.91,
|
34515 |
+
"learning_rate": 4.676850029675444e-06,
|
34516 |
+
"loss": 0.0487,
|
34517 |
+
"step": 170430
|
34518 |
+
},
|
34519 |
+
{
|
34520 |
+
"epoch": 47.92,
|
34521 |
+
"learning_rate": 4.6581076437697185e-06,
|
34522 |
+
"loss": 0.0376,
|
34523 |
+
"step": 170460
|
34524 |
+
},
|
34525 |
+
{
|
34526 |
+
"epoch": 47.93,
|
34527 |
+
"learning_rate": 4.639365257863993e-06,
|
34528 |
+
"loss": 0.0466,
|
34529 |
+
"step": 170490
|
34530 |
+
},
|
34531 |
+
{
|
34532 |
+
"epoch": 47.94,
|
34533 |
+
"learning_rate": 4.620622871958267e-06,
|
34534 |
+
"loss": 0.0386,
|
34535 |
+
"step": 170520
|
34536 |
+
},
|
34537 |
+
{
|
34538 |
+
"epoch": 47.95,
|
34539 |
+
"learning_rate": 4.601880486052541e-06,
|
34540 |
+
"loss": 0.0408,
|
34541 |
+
"step": 170550
|
34542 |
+
},
|
34543 |
+
{
|
34544 |
+
"epoch": 47.96,
|
34545 |
+
"learning_rate": 4.583138100146816e-06,
|
34546 |
+
"loss": 0.0453,
|
34547 |
+
"step": 170580
|
34548 |
+
},
|
34549 |
+
{
|
34550 |
+
"epoch": 47.96,
|
34551 |
+
"learning_rate": 4.5643957142410895e-06,
|
34552 |
+
"loss": 0.0374,
|
34553 |
+
"step": 170610
|
34554 |
+
},
|
34555 |
+
{
|
34556 |
+
"epoch": 47.97,
|
34557 |
+
"learning_rate": 4.5456533283353645e-06,
|
34558 |
+
"loss": 0.0468,
|
34559 |
+
"step": 170640
|
34560 |
+
},
|
34561 |
+
{
|
34562 |
+
"epoch": 47.98,
|
34563 |
+
"learning_rate": 4.526910942429638e-06,
|
34564 |
+
"loss": 0.0421,
|
34565 |
+
"step": 170670
|
34566 |
+
},
|
34567 |
+
{
|
34568 |
+
"epoch": 47.99,
|
34569 |
+
"learning_rate": 4.508168556523912e-06,
|
34570 |
+
"loss": 0.043,
|
34571 |
+
"step": 170700
|
34572 |
+
},
|
34573 |
+
{
|
34574 |
+
"epoch": 48.0,
|
34575 |
+
"learning_rate": 4.490050916815044e-06,
|
34576 |
+
"loss": 0.0501,
|
34577 |
+
"step": 170730
|
34578 |
+
},
|
34579 |
+
{
|
34580 |
+
"epoch": 48.0,
|
34581 |
+
"eval_loss": 0.10937908291816711,
|
34582 |
+
"eval_runtime": 643.5663,
|
34583 |
+
"eval_samples_per_second": 24.913,
|
34584 |
+
"eval_steps_per_second": 0.78,
|
34585 |
+
"eval_wer": 0.09180052679704666,
|
34586 |
+
"step": 170736
|
34587 |
+
},
|
34588 |
+
{
|
34589 |
+
"epoch": 48.01,
|
34590 |
+
"learning_rate": 4.471308530909318e-06,
|
34591 |
+
"loss": 0.0404,
|
34592 |
+
"step": 170760
|
34593 |
+
},
|
34594 |
+
{
|
34595 |
+
"epoch": 48.02,
|
34596 |
+
"learning_rate": 4.452566145003593e-06,
|
34597 |
+
"loss": 0.0433,
|
34598 |
+
"step": 170790
|
34599 |
+
},
|
34600 |
+
{
|
34601 |
+
"epoch": 48.02,
|
34602 |
+
"learning_rate": 4.433823759097866e-06,
|
34603 |
+
"loss": 0.0372,
|
34604 |
+
"step": 170820
|
34605 |
+
},
|
34606 |
+
{
|
34607 |
+
"epoch": 48.03,
|
34608 |
+
"learning_rate": 4.415081373192141e-06,
|
34609 |
+
"loss": 0.0425,
|
34610 |
+
"step": 170850
|
34611 |
+
},
|
34612 |
+
{
|
34613 |
+
"epoch": 48.04,
|
34614 |
+
"learning_rate": 4.396338987286415e-06,
|
34615 |
+
"loss": 0.0418,
|
34616 |
+
"step": 170880
|
34617 |
+
},
|
34618 |
+
{
|
34619 |
+
"epoch": 48.05,
|
34620 |
+
"learning_rate": 4.37759660138069e-06,
|
34621 |
+
"loss": 0.0448,
|
34622 |
+
"step": 170910
|
34623 |
+
},
|
34624 |
+
{
|
34625 |
+
"epoch": 48.06,
|
34626 |
+
"learning_rate": 4.358854215474963e-06,
|
34627 |
+
"loss": 0.0454,
|
34628 |
+
"step": 170940
|
34629 |
+
},
|
34630 |
+
{
|
34631 |
+
"epoch": 48.07,
|
34632 |
+
"learning_rate": 4.340111829569238e-06,
|
34633 |
+
"loss": 0.0397,
|
34634 |
+
"step": 170970
|
34635 |
+
},
|
34636 |
+
{
|
34637 |
+
"epoch": 48.07,
|
34638 |
+
"learning_rate": 4.3213694436635114e-06,
|
34639 |
+
"loss": 0.0473,
|
34640 |
+
"step": 171000
|
34641 |
+
},
|
34642 |
+
{
|
34643 |
+
"epoch": 48.08,
|
34644 |
+
"learning_rate": 4.3026270577577865e-06,
|
34645 |
+
"loss": 0.0426,
|
34646 |
+
"step": 171030
|
34647 |
+
},
|
34648 |
+
{
|
34649 |
+
"epoch": 48.09,
|
34650 |
+
"learning_rate": 4.283884671852061e-06,
|
34651 |
+
"loss": 0.045,
|
34652 |
+
"step": 171060
|
34653 |
+
},
|
34654 |
+
{
|
34655 |
+
"epoch": 48.1,
|
34656 |
+
"learning_rate": 4.265142285946334e-06,
|
34657 |
+
"loss": 0.0452,
|
34658 |
+
"step": 171090
|
34659 |
+
},
|
34660 |
+
{
|
34661 |
+
"epoch": 48.11,
|
34662 |
+
"learning_rate": 4.246399900040609e-06,
|
34663 |
+
"loss": 0.0401,
|
34664 |
+
"step": 171120
|
34665 |
+
},
|
34666 |
+
{
|
34667 |
+
"epoch": 48.12,
|
34668 |
+
"learning_rate": 4.227657514134882e-06,
|
34669 |
+
"loss": 0.0424,
|
34670 |
+
"step": 171150
|
34671 |
+
},
|
34672 |
+
{
|
34673 |
+
"epoch": 48.12,
|
34674 |
+
"learning_rate": 4.2089151282291574e-06,
|
34675 |
+
"loss": 0.0393,
|
34676 |
+
"step": 171180
|
34677 |
+
},
|
34678 |
+
{
|
34679 |
+
"epoch": 48.13,
|
34680 |
+
"learning_rate": 4.190172742323431e-06,
|
34681 |
+
"loss": 0.0406,
|
34682 |
+
"step": 171210
|
34683 |
+
},
|
34684 |
+
{
|
34685 |
+
"epoch": 48.14,
|
34686 |
+
"learning_rate": 4.171430356417706e-06,
|
34687 |
+
"loss": 0.0445,
|
34688 |
+
"step": 171240
|
34689 |
+
},
|
34690 |
+
{
|
34691 |
+
"epoch": 48.15,
|
34692 |
+
"learning_rate": 4.15268797051198e-06,
|
34693 |
+
"loss": 0.039,
|
34694 |
+
"step": 171270
|
34695 |
+
},
|
34696 |
+
{
|
34697 |
+
"epoch": 48.16,
|
34698 |
+
"learning_rate": 4.133945584606254e-06,
|
34699 |
+
"loss": 0.0456,
|
34700 |
+
"step": 171300
|
34701 |
+
},
|
34702 |
+
{
|
34703 |
+
"epoch": 48.17,
|
34704 |
+
"learning_rate": 4.115203198700528e-06,
|
34705 |
+
"loss": 0.0429,
|
34706 |
+
"step": 171330
|
34707 |
+
},
|
34708 |
+
{
|
34709 |
+
"epoch": 48.18,
|
34710 |
+
"learning_rate": 4.096460812794803e-06,
|
34711 |
+
"loss": 0.0413,
|
34712 |
+
"step": 171360
|
34713 |
+
},
|
34714 |
+
{
|
34715 |
+
"epoch": 48.18,
|
34716 |
+
"learning_rate": 4.077718426889077e-06,
|
34717 |
+
"loss": 0.0466,
|
34718 |
+
"step": 171390
|
34719 |
+
},
|
34720 |
+
{
|
34721 |
+
"epoch": 48.19,
|
34722 |
+
"learning_rate": 4.05897604098335e-06,
|
34723 |
+
"loss": 0.0402,
|
34724 |
+
"step": 171420
|
34725 |
+
},
|
34726 |
+
{
|
34727 |
+
"epoch": 48.2,
|
34728 |
+
"learning_rate": 4.040233655077625e-06,
|
34729 |
+
"loss": 0.0428,
|
34730 |
+
"step": 171450
|
34731 |
+
},
|
34732 |
+
{
|
34733 |
+
"epoch": 48.21,
|
34734 |
+
"learning_rate": 4.0214912691718985e-06,
|
34735 |
+
"loss": 0.042,
|
34736 |
+
"step": 171480
|
34737 |
+
},
|
34738 |
+
{
|
34739 |
+
"epoch": 48.22,
|
34740 |
+
"learning_rate": 4.0027488832661735e-06,
|
34741 |
+
"loss": 0.0388,
|
34742 |
+
"step": 171510
|
34743 |
+
},
|
34744 |
+
{
|
34745 |
+
"epoch": 48.23,
|
34746 |
+
"learning_rate": 3.984006497360448e-06,
|
34747 |
+
"loss": 0.048,
|
34748 |
+
"step": 171540
|
34749 |
+
},
|
34750 |
+
{
|
34751 |
+
"epoch": 48.23,
|
34752 |
+
"learning_rate": 3.965264111454722e-06,
|
34753 |
+
"loss": 0.0399,
|
34754 |
+
"step": 171570
|
34755 |
+
},
|
34756 |
+
{
|
34757 |
+
"epoch": 48.24,
|
34758 |
+
"learning_rate": 3.946521725548996e-06,
|
34759 |
+
"loss": 0.0477,
|
34760 |
+
"step": 171600
|
34761 |
+
},
|
34762 |
+
{
|
34763 |
+
"epoch": 48.25,
|
34764 |
+
"learning_rate": 3.92777933964327e-06,
|
34765 |
+
"loss": 0.0384,
|
34766 |
+
"step": 171630
|
34767 |
+
},
|
34768 |
+
{
|
34769 |
+
"epoch": 48.26,
|
34770 |
+
"learning_rate": 3.9090369537375445e-06,
|
34771 |
+
"loss": 0.0415,
|
34772 |
+
"step": 171660
|
34773 |
+
},
|
34774 |
+
{
|
34775 |
+
"epoch": 48.27,
|
34776 |
+
"learning_rate": 3.890294567831818e-06,
|
34777 |
+
"loss": 0.0421,
|
34778 |
+
"step": 171690
|
34779 |
+
},
|
34780 |
+
{
|
34781 |
+
"epoch": 48.28,
|
34782 |
+
"learning_rate": 3.871552181926093e-06,
|
34783 |
+
"loss": 0.041,
|
34784 |
+
"step": 171720
|
34785 |
+
},
|
34786 |
+
{
|
34787 |
+
"epoch": 48.29,
|
34788 |
+
"learning_rate": 3.852809796020366e-06,
|
34789 |
+
"loss": 0.0445,
|
34790 |
+
"step": 171750
|
34791 |
+
},
|
34792 |
+
{
|
34793 |
+
"epoch": 48.29,
|
34794 |
+
"learning_rate": 3.834067410114641e-06,
|
34795 |
+
"loss": 0.0414,
|
34796 |
+
"step": 171780
|
34797 |
+
},
|
34798 |
+
{
|
34799 |
+
"epoch": 48.3,
|
34800 |
+
"learning_rate": 3.8153250242089154e-06,
|
34801 |
+
"loss": 0.0386,
|
34802 |
+
"step": 171810
|
34803 |
+
},
|
34804 |
+
{
|
34805 |
+
"epoch": 48.31,
|
34806 |
+
"learning_rate": 3.796582638303189e-06,
|
34807 |
+
"loss": 0.0455,
|
34808 |
+
"step": 171840
|
34809 |
+
},
|
34810 |
+
{
|
34811 |
+
"epoch": 48.32,
|
34812 |
+
"learning_rate": 3.777840252397464e-06,
|
34813 |
+
"loss": 0.0372,
|
34814 |
+
"step": 171870
|
34815 |
+
},
|
34816 |
+
{
|
34817 |
+
"epoch": 48.33,
|
34818 |
+
"learning_rate": 3.7590978664917376e-06,
|
34819 |
+
"loss": 0.043,
|
34820 |
+
"step": 171900
|
34821 |
+
},
|
34822 |
+
{
|
34823 |
+
"epoch": 48.34,
|
34824 |
+
"learning_rate": 3.740355480586012e-06,
|
34825 |
+
"loss": 0.042,
|
34826 |
+
"step": 171930
|
34827 |
+
},
|
34828 |
+
{
|
34829 |
+
"epoch": 48.34,
|
34830 |
+
"learning_rate": 3.721613094680286e-06,
|
34831 |
+
"loss": 0.0422,
|
34832 |
+
"step": 171960
|
34833 |
+
},
|
34834 |
+
{
|
34835 |
+
"epoch": 48.35,
|
34836 |
+
"learning_rate": 3.7028707087745606e-06,
|
34837 |
+
"loss": 0.045,
|
34838 |
+
"step": 171990
|
34839 |
+
},
|
34840 |
+
{
|
34841 |
+
"epoch": 48.36,
|
34842 |
+
"learning_rate": 3.684128322868835e-06,
|
34843 |
+
"loss": 0.0401,
|
34844 |
+
"step": 172020
|
34845 |
+
},
|
34846 |
+
{
|
34847 |
+
"epoch": 48.37,
|
34848 |
+
"learning_rate": 3.665385936963109e-06,
|
34849 |
+
"loss": 0.0422,
|
34850 |
+
"step": 172050
|
34851 |
+
},
|
34852 |
+
{
|
34853 |
+
"epoch": 48.38,
|
34854 |
+
"learning_rate": 3.646643551057383e-06,
|
34855 |
+
"loss": 0.0401,
|
34856 |
+
"step": 172080
|
34857 |
+
},
|
34858 |
+
{
|
34859 |
+
"epoch": 48.39,
|
34860 |
+
"learning_rate": 3.6279011651516573e-06,
|
34861 |
+
"loss": 0.0428,
|
34862 |
+
"step": 172110
|
34863 |
+
},
|
34864 |
+
{
|
34865 |
+
"epoch": 48.39,
|
34866 |
+
"learning_rate": 3.6091587792459315e-06,
|
34867 |
+
"loss": 0.0469,
|
34868 |
+
"step": 172140
|
34869 |
+
},
|
34870 |
+
{
|
34871 |
+
"epoch": 48.4,
|
34872 |
+
"learning_rate": 3.5904163933402053e-06,
|
34873 |
+
"loss": 0.0406,
|
34874 |
+
"step": 172170
|
34875 |
+
},
|
34876 |
+
{
|
34877 |
+
"epoch": 48.41,
|
34878 |
+
"learning_rate": 3.57167400743448e-06,
|
34879 |
+
"loss": 0.0427,
|
34880 |
+
"step": 172200
|
34881 |
+
},
|
34882 |
+
{
|
34883 |
+
"epoch": 48.42,
|
34884 |
+
"learning_rate": 3.5529316215287537e-06,
|
34885 |
+
"loss": 0.0421,
|
34886 |
+
"step": 172230
|
34887 |
+
},
|
34888 |
+
{
|
34889 |
+
"epoch": 48.43,
|
34890 |
+
"learning_rate": 3.5341892356230283e-06,
|
34891 |
+
"loss": 0.0402,
|
34892 |
+
"step": 172260
|
34893 |
+
},
|
34894 |
+
{
|
34895 |
+
"epoch": 48.44,
|
34896 |
+
"learning_rate": 3.515446849717303e-06,
|
34897 |
+
"loss": 0.0505,
|
34898 |
+
"step": 172290
|
34899 |
+
},
|
34900 |
+
{
|
34901 |
+
"epoch": 48.45,
|
34902 |
+
"learning_rate": 3.4967044638115767e-06,
|
34903 |
+
"loss": 0.0371,
|
34904 |
+
"step": 172320
|
34905 |
+
},
|
34906 |
+
{
|
34907 |
+
"epoch": 48.45,
|
34908 |
+
"learning_rate": 3.4779620779058513e-06,
|
34909 |
+
"loss": 0.0437,
|
34910 |
+
"step": 172350
|
34911 |
+
},
|
34912 |
+
{
|
34913 |
+
"epoch": 48.46,
|
34914 |
+
"learning_rate": 3.459219692000125e-06,
|
34915 |
+
"loss": 0.0422,
|
34916 |
+
"step": 172380
|
34917 |
+
},
|
34918 |
+
{
|
34919 |
+
"epoch": 48.47,
|
34920 |
+
"learning_rate": 3.4404773060943993e-06,
|
34921 |
+
"loss": 0.0442,
|
34922 |
+
"step": 172410
|
34923 |
+
},
|
34924 |
+
{
|
34925 |
+
"epoch": 48.48,
|
34926 |
+
"learning_rate": 3.421734920188673e-06,
|
34927 |
+
"loss": 0.0437,
|
34928 |
+
"step": 172440
|
34929 |
+
},
|
34930 |
+
{
|
34931 |
+
"epoch": 48.49,
|
34932 |
+
"learning_rate": 3.4029925342829476e-06,
|
34933 |
+
"loss": 0.0351,
|
34934 |
+
"step": 172470
|
34935 |
+
},
|
34936 |
+
{
|
34937 |
+
"epoch": 48.5,
|
34938 |
+
"learning_rate": 3.3842501483772223e-06,
|
34939 |
+
"loss": 0.0466,
|
34940 |
+
"step": 172500
|
34941 |
+
},
|
34942 |
+
{
|
34943 |
+
"epoch": 48.5,
|
34944 |
+
"learning_rate": 3.365507762471496e-06,
|
34945 |
+
"loss": 0.0441,
|
34946 |
+
"step": 172530
|
34947 |
+
},
|
34948 |
+
{
|
34949 |
+
"epoch": 48.51,
|
34950 |
+
"learning_rate": 3.3467653765657706e-06,
|
34951 |
+
"loss": 0.0428,
|
34952 |
+
"step": 172560
|
34953 |
+
},
|
34954 |
+
{
|
34955 |
+
"epoch": 48.52,
|
34956 |
+
"learning_rate": 3.3280229906600444e-06,
|
34957 |
+
"loss": 0.0438,
|
34958 |
+
"step": 172590
|
34959 |
+
},
|
34960 |
+
{
|
34961 |
+
"epoch": 48.53,
|
34962 |
+
"learning_rate": 3.309280604754319e-06,
|
34963 |
+
"loss": 0.0371,
|
34964 |
+
"step": 172620
|
34965 |
+
},
|
34966 |
+
{
|
34967 |
+
"epoch": 48.54,
|
34968 |
+
"learning_rate": 3.2905382188485928e-06,
|
34969 |
+
"loss": 0.0445,
|
34970 |
+
"step": 172650
|
34971 |
+
},
|
34972 |
+
{
|
34973 |
+
"epoch": 48.55,
|
34974 |
+
"learning_rate": 3.271795832942867e-06,
|
34975 |
+
"loss": 0.041,
|
34976 |
+
"step": 172680
|
34977 |
+
},
|
34978 |
+
{
|
34979 |
+
"epoch": 48.55,
|
34980 |
+
"learning_rate": 3.253053447037141e-06,
|
34981 |
+
"loss": 0.0432,
|
34982 |
+
"step": 172710
|
34983 |
+
},
|
34984 |
+
{
|
34985 |
+
"epoch": 48.56,
|
34986 |
+
"learning_rate": 3.2343110611314154e-06,
|
34987 |
+
"loss": 0.045,
|
34988 |
+
"step": 172740
|
34989 |
+
},
|
34990 |
+
{
|
34991 |
+
"epoch": 48.57,
|
34992 |
+
"learning_rate": 3.21556867522569e-06,
|
34993 |
+
"loss": 0.0358,
|
34994 |
+
"step": 172770
|
34995 |
+
},
|
34996 |
+
{
|
34997 |
+
"epoch": 48.58,
|
34998 |
+
"learning_rate": 3.1968262893199637e-06,
|
34999 |
+
"loss": 0.0449,
|
35000 |
+
"step": 172800
|
35001 |
+
},
|
35002 |
+
{
|
35003 |
+
"epoch": 48.59,
|
35004 |
+
"learning_rate": 3.1780839034142383e-06,
|
35005 |
+
"loss": 0.0411,
|
35006 |
+
"step": 172830
|
35007 |
+
},
|
35008 |
+
{
|
35009 |
+
"epoch": 48.6,
|
35010 |
+
"learning_rate": 3.159341517508512e-06,
|
35011 |
+
"loss": 0.0416,
|
35012 |
+
"step": 172860
|
35013 |
+
},
|
35014 |
+
{
|
35015 |
+
"epoch": 48.61,
|
35016 |
+
"learning_rate": 3.1405991316027867e-06,
|
35017 |
+
"loss": 0.0448,
|
35018 |
+
"step": 172890
|
35019 |
+
},
|
35020 |
+
{
|
35021 |
+
"epoch": 48.61,
|
35022 |
+
"learning_rate": 3.121856745697061e-06,
|
35023 |
+
"loss": 0.0381,
|
35024 |
+
"step": 172920
|
35025 |
+
},
|
35026 |
+
{
|
35027 |
+
"epoch": 48.62,
|
35028 |
+
"learning_rate": 3.103114359791335e-06,
|
35029 |
+
"loss": 0.0445,
|
35030 |
+
"step": 172950
|
35031 |
+
},
|
35032 |
+
{
|
35033 |
+
"epoch": 48.63,
|
35034 |
+
"learning_rate": 3.0843719738856093e-06,
|
35035 |
+
"loss": 0.0413,
|
35036 |
+
"step": 172980
|
35037 |
+
},
|
35038 |
+
{
|
35039 |
+
"epoch": 48.64,
|
35040 |
+
"learning_rate": 3.065629587979883e-06,
|
35041 |
+
"loss": 0.0426,
|
35042 |
+
"step": 173010
|
35043 |
+
},
|
35044 |
+
{
|
35045 |
+
"epoch": 48.65,
|
35046 |
+
"learning_rate": 3.0468872020741573e-06,
|
35047 |
+
"loss": 0.0466,
|
35048 |
+
"step": 173040
|
35049 |
+
},
|
35050 |
+
{
|
35051 |
+
"epoch": 48.66,
|
35052 |
+
"learning_rate": 3.0281448161684315e-06,
|
35053 |
+
"loss": 0.0397,
|
35054 |
+
"step": 173070
|
35055 |
+
},
|
35056 |
+
{
|
35057 |
+
"epoch": 48.66,
|
35058 |
+
"learning_rate": 3.009402430262706e-06,
|
35059 |
+
"loss": 0.0432,
|
35060 |
+
"step": 173100
|
35061 |
+
},
|
35062 |
+
{
|
35063 |
+
"epoch": 48.67,
|
35064 |
+
"learning_rate": 2.9906600443569803e-06,
|
35065 |
+
"loss": 0.0419,
|
35066 |
+
"step": 173130
|
35067 |
+
},
|
35068 |
+
{
|
35069 |
+
"epoch": 48.68,
|
35070 |
+
"learning_rate": 2.9719176584512544e-06,
|
35071 |
+
"loss": 0.0421,
|
35072 |
+
"step": 173160
|
35073 |
+
},
|
35074 |
+
{
|
35075 |
+
"epoch": 48.69,
|
35076 |
+
"learning_rate": 2.9531752725455286e-06,
|
35077 |
+
"loss": 0.0457,
|
35078 |
+
"step": 173190
|
35079 |
+
},
|
35080 |
+
{
|
35081 |
+
"epoch": 48.7,
|
35082 |
+
"learning_rate": 2.934432886639803e-06,
|
35083 |
+
"loss": 0.0388,
|
35084 |
+
"step": 173220
|
35085 |
+
},
|
35086 |
+
{
|
35087 |
+
"epoch": 48.71,
|
35088 |
+
"learning_rate": 2.915690500734077e-06,
|
35089 |
+
"loss": 0.0451,
|
35090 |
+
"step": 173250
|
35091 |
+
},
|
35092 |
+
{
|
35093 |
+
"epoch": 48.72,
|
35094 |
+
"learning_rate": 2.896948114828351e-06,
|
35095 |
+
"loss": 0.0412,
|
35096 |
+
"step": 173280
|
35097 |
+
},
|
35098 |
+
{
|
35099 |
+
"epoch": 48.72,
|
35100 |
+
"learning_rate": 2.878205728922625e-06,
|
35101 |
+
"loss": 0.0411,
|
35102 |
+
"step": 173310
|
35103 |
+
},
|
35104 |
+
{
|
35105 |
+
"epoch": 48.73,
|
35106 |
+
"learning_rate": 2.860088089213757e-06,
|
35107 |
+
"loss": 0.0485,
|
35108 |
+
"step": 173340
|
35109 |
+
},
|
35110 |
+
{
|
35111 |
+
"epoch": 48.74,
|
35112 |
+
"learning_rate": 2.8413457033080313e-06,
|
35113 |
+
"loss": 0.0404,
|
35114 |
+
"step": 173370
|
35115 |
+
},
|
35116 |
+
{
|
35117 |
+
"epoch": 48.75,
|
35118 |
+
"learning_rate": 2.8226033174023054e-06,
|
35119 |
+
"loss": 0.0442,
|
35120 |
+
"step": 173400
|
35121 |
+
},
|
35122 |
+
{
|
35123 |
+
"epoch": 48.76,
|
35124 |
+
"learning_rate": 2.8038609314965796e-06,
|
35125 |
+
"loss": 0.0409,
|
35126 |
+
"step": 173430
|
35127 |
+
},
|
35128 |
+
{
|
35129 |
+
"epoch": 48.77,
|
35130 |
+
"learning_rate": 2.785118545590854e-06,
|
35131 |
+
"loss": 0.0418,
|
35132 |
+
"step": 173460
|
35133 |
+
},
|
35134 |
+
{
|
35135 |
+
"epoch": 48.77,
|
35136 |
+
"learning_rate": 2.766376159685128e-06,
|
35137 |
+
"loss": 0.045,
|
35138 |
+
"step": 173490
|
35139 |
+
},
|
35140 |
+
{
|
35141 |
+
"epoch": 48.78,
|
35142 |
+
"learning_rate": 2.747633773779402e-06,
|
35143 |
+
"loss": 0.0374,
|
35144 |
+
"step": 173520
|
35145 |
+
},
|
35146 |
+
{
|
35147 |
+
"epoch": 48.79,
|
35148 |
+
"learning_rate": 2.7288913878736764e-06,
|
35149 |
+
"loss": 0.0434,
|
35150 |
+
"step": 173550
|
35151 |
+
},
|
35152 |
+
{
|
35153 |
+
"epoch": 48.8,
|
35154 |
+
"learning_rate": 2.7101490019679506e-06,
|
35155 |
+
"loss": 0.0414,
|
35156 |
+
"step": 173580
|
35157 |
+
},
|
35158 |
+
{
|
35159 |
+
"epoch": 48.81,
|
35160 |
+
"learning_rate": 2.6914066160622248e-06,
|
35161 |
+
"loss": 0.0424,
|
35162 |
+
"step": 173610
|
35163 |
+
},
|
35164 |
+
{
|
35165 |
+
"epoch": 48.82,
|
35166 |
+
"learning_rate": 2.672664230156499e-06,
|
35167 |
+
"loss": 0.0449,
|
35168 |
+
"step": 173640
|
35169 |
+
},
|
35170 |
+
{
|
35171 |
+
"epoch": 48.82,
|
35172 |
+
"learning_rate": 2.653921844250773e-06,
|
35173 |
+
"loss": 0.0411,
|
35174 |
+
"step": 173670
|
35175 |
+
},
|
35176 |
+
{
|
35177 |
+
"epoch": 48.83,
|
35178 |
+
"learning_rate": 2.6351794583450474e-06,
|
35179 |
+
"loss": 0.0442,
|
35180 |
+
"step": 173700
|
35181 |
+
},
|
35182 |
+
{
|
35183 |
+
"epoch": 48.84,
|
35184 |
+
"learning_rate": 2.6164370724393215e-06,
|
35185 |
+
"loss": 0.045,
|
35186 |
+
"step": 173730
|
35187 |
+
},
|
35188 |
+
{
|
35189 |
+
"epoch": 48.85,
|
35190 |
+
"learning_rate": 2.5976946865335957e-06,
|
35191 |
+
"loss": 0.0415,
|
35192 |
+
"step": 173760
|
35193 |
+
},
|
35194 |
+
{
|
35195 |
+
"epoch": 48.86,
|
35196 |
+
"learning_rate": 2.57895230062787e-06,
|
35197 |
+
"loss": 0.045,
|
35198 |
+
"step": 173790
|
35199 |
+
},
|
35200 |
+
{
|
35201 |
+
"epoch": 48.87,
|
35202 |
+
"learning_rate": 2.560209914722144e-06,
|
35203 |
+
"loss": 0.037,
|
35204 |
+
"step": 173820
|
35205 |
+
},
|
35206 |
+
{
|
35207 |
+
"epoch": 48.88,
|
35208 |
+
"learning_rate": 2.5414675288164187e-06,
|
35209 |
+
"loss": 0.0469,
|
35210 |
+
"step": 173850
|
35211 |
+
},
|
35212 |
+
{
|
35213 |
+
"epoch": 48.88,
|
35214 |
+
"learning_rate": 2.522725142910693e-06,
|
35215 |
+
"loss": 0.0386,
|
35216 |
+
"step": 173880
|
35217 |
+
},
|
35218 |
+
{
|
35219 |
+
"epoch": 48.89,
|
35220 |
+
"learning_rate": 2.503982757004967e-06,
|
35221 |
+
"loss": 0.0422,
|
35222 |
+
"step": 173910
|
35223 |
+
},
|
35224 |
+
{
|
35225 |
+
"epoch": 48.9,
|
35226 |
+
"learning_rate": 2.485240371099241e-06,
|
35227 |
+
"loss": 0.0428,
|
35228 |
+
"step": 173940
|
35229 |
+
},
|
35230 |
+
{
|
35231 |
+
"epoch": 48.91,
|
35232 |
+
"learning_rate": 2.466497985193515e-06,
|
35233 |
+
"loss": 0.0381,
|
35234 |
+
"step": 173970
|
35235 |
+
},
|
35236 |
+
{
|
35237 |
+
"epoch": 48.92,
|
35238 |
+
"learning_rate": 2.4477555992877893e-06,
|
35239 |
+
"loss": 0.0446,
|
35240 |
+
"step": 174000
|
35241 |
+
},
|
35242 |
+
{
|
35243 |
+
"epoch": 48.93,
|
35244 |
+
"learning_rate": 2.4290132133820634e-06,
|
35245 |
+
"loss": 0.0403,
|
35246 |
+
"step": 174030
|
35247 |
+
},
|
35248 |
+
{
|
35249 |
+
"epoch": 48.93,
|
35250 |
+
"learning_rate": 2.4102708274763376e-06,
|
35251 |
+
"loss": 0.0417,
|
35252 |
+
"step": 174060
|
35253 |
+
},
|
35254 |
+
{
|
35255 |
+
"epoch": 48.94,
|
35256 |
+
"learning_rate": 2.3915284415706123e-06,
|
35257 |
+
"loss": 0.0428,
|
35258 |
+
"step": 174090
|
35259 |
+
},
|
35260 |
+
{
|
35261 |
+
"epoch": 48.95,
|
35262 |
+
"learning_rate": 2.3727860556648864e-06,
|
35263 |
+
"loss": 0.0367,
|
35264 |
+
"step": 174120
|
35265 |
+
},
|
35266 |
+
{
|
35267 |
+
"epoch": 48.96,
|
35268 |
+
"learning_rate": 2.3540436697591606e-06,
|
35269 |
+
"loss": 0.0465,
|
35270 |
+
"step": 174150
|
35271 |
+
},
|
35272 |
+
{
|
35273 |
+
"epoch": 48.97,
|
35274 |
+
"learning_rate": 2.335301283853435e-06,
|
35275 |
+
"loss": 0.0397,
|
35276 |
+
"step": 174180
|
35277 |
+
},
|
35278 |
+
{
|
35279 |
+
"epoch": 48.98,
|
35280 |
+
"learning_rate": 2.316558897947709e-06,
|
35281 |
+
"loss": 0.0429,
|
35282 |
+
"step": 174210
|
35283 |
+
},
|
35284 |
+
{
|
35285 |
+
"epoch": 48.98,
|
35286 |
+
"learning_rate": 2.2978165120419828e-06,
|
35287 |
+
"loss": 0.0472,
|
35288 |
+
"step": 174240
|
35289 |
+
},
|
35290 |
+
{
|
35291 |
+
"epoch": 48.99,
|
35292 |
+
"learning_rate": 2.279074126136257e-06,
|
35293 |
+
"loss": 0.0388,
|
35294 |
+
"step": 174270
|
35295 |
+
},
|
35296 |
+
{
|
35297 |
+
"epoch": 49.0,
|
35298 |
+
"eval_loss": 0.10990184545516968,
|
35299 |
+
"eval_runtime": 658.3127,
|
35300 |
+
"eval_samples_per_second": 24.355,
|
35301 |
+
"eval_steps_per_second": 0.763,
|
35302 |
+
"eval_wer": 0.09180735059298786,
|
35303 |
+
"step": 174293
|
35304 |
+
},
|
35305 |
+
{
|
35306 |
+
"epoch": 49.0,
|
35307 |
+
"learning_rate": 2.260331740230531e-06,
|
35308 |
+
"loss": 0.0459,
|
35309 |
+
"step": 174300
|
35310 |
+
},
|
35311 |
+
{
|
35312 |
+
"epoch": 49.01,
|
35313 |
+
"learning_rate": 2.2415893543248058e-06,
|
35314 |
+
"loss": 0.0399,
|
35315 |
+
"step": 174330
|
35316 |
+
},
|
35317 |
+
{
|
35318 |
+
"epoch": 49.02,
|
35319 |
+
"learning_rate": 2.22284696841908e-06,
|
35320 |
+
"loss": 0.0474,
|
35321 |
+
"step": 174360
|
35322 |
+
},
|
35323 |
+
{
|
35324 |
+
"epoch": 49.03,
|
35325 |
+
"learning_rate": 2.204104582513354e-06,
|
35326 |
+
"loss": 0.04,
|
35327 |
+
"step": 174390
|
35328 |
+
},
|
35329 |
+
{
|
35330 |
+
"epoch": 49.04,
|
35331 |
+
"learning_rate": 2.1853621966076283e-06,
|
35332 |
+
"loss": 0.0368,
|
35333 |
+
"step": 174420
|
35334 |
+
},
|
35335 |
+
{
|
35336 |
+
"epoch": 49.04,
|
35337 |
+
"learning_rate": 2.1666198107019025e-06,
|
35338 |
+
"loss": 0.0473,
|
35339 |
+
"step": 174450
|
35340 |
+
},
|
35341 |
+
{
|
35342 |
+
"epoch": 49.05,
|
35343 |
+
"learning_rate": 2.1478774247961767e-06,
|
35344 |
+
"loss": 0.0403,
|
35345 |
+
"step": 174480
|
35346 |
+
},
|
35347 |
+
{
|
35348 |
+
"epoch": 49.06,
|
35349 |
+
"learning_rate": 2.129135038890451e-06,
|
35350 |
+
"loss": 0.0454,
|
35351 |
+
"step": 174510
|
35352 |
+
},
|
35353 |
+
{
|
35354 |
+
"epoch": 49.07,
|
35355 |
+
"learning_rate": 2.1103926529847247e-06,
|
35356 |
+
"loss": 0.0436,
|
35357 |
+
"step": 174540
|
35358 |
+
},
|
35359 |
+
{
|
35360 |
+
"epoch": 49.08,
|
35361 |
+
"learning_rate": 2.0916502670789993e-06,
|
35362 |
+
"loss": 0.0431,
|
35363 |
+
"step": 174570
|
35364 |
+
},
|
35365 |
+
{
|
35366 |
+
"epoch": 49.09,
|
35367 |
+
"learning_rate": 2.0729078811732735e-06,
|
35368 |
+
"loss": 0.0444,
|
35369 |
+
"step": 174600
|
35370 |
+
},
|
35371 |
+
{
|
35372 |
+
"epoch": 49.09,
|
35373 |
+
"learning_rate": 2.0541654952675477e-06,
|
35374 |
+
"loss": 0.0378,
|
35375 |
+
"step": 174630
|
35376 |
+
},
|
35377 |
+
{
|
35378 |
+
"epoch": 49.1,
|
35379 |
+
"learning_rate": 2.035423109361822e-06,
|
35380 |
+
"loss": 0.0445,
|
35381 |
+
"step": 174660
|
35382 |
+
},
|
35383 |
+
{
|
35384 |
+
"epoch": 49.11,
|
35385 |
+
"learning_rate": 2.016680723456096e-06,
|
35386 |
+
"loss": 0.0444,
|
35387 |
+
"step": 174690
|
35388 |
+
},
|
35389 |
+
{
|
35390 |
+
"epoch": 49.12,
|
35391 |
+
"learning_rate": 1.9979383375503703e-06,
|
35392 |
+
"loss": 0.0386,
|
35393 |
+
"step": 174720
|
35394 |
+
},
|
35395 |
+
{
|
35396 |
+
"epoch": 49.13,
|
35397 |
+
"learning_rate": 1.9791959516446444e-06,
|
35398 |
+
"loss": 0.0432,
|
35399 |
+
"step": 174750
|
35400 |
+
},
|
35401 |
+
{
|
35402 |
+
"epoch": 49.14,
|
35403 |
+
"learning_rate": 1.9604535657389186e-06,
|
35404 |
+
"loss": 0.0397,
|
35405 |
+
"step": 174780
|
35406 |
+
},
|
35407 |
+
{
|
35408 |
+
"epoch": 49.15,
|
35409 |
+
"learning_rate": 1.941711179833193e-06,
|
35410 |
+
"loss": 0.0444,
|
35411 |
+
"step": 174810
|
35412 |
+
},
|
35413 |
+
{
|
35414 |
+
"epoch": 49.15,
|
35415 |
+
"learning_rate": 1.922968793927467e-06,
|
35416 |
+
"loss": 0.0438,
|
35417 |
+
"step": 174840
|
35418 |
+
},
|
35419 |
+
{
|
35420 |
+
"epoch": 49.16,
|
35421 |
+
"learning_rate": 1.9042264080217414e-06,
|
35422 |
+
"loss": 0.0414,
|
35423 |
+
"step": 174870
|
35424 |
+
},
|
35425 |
+
{
|
35426 |
+
"epoch": 49.17,
|
35427 |
+
"learning_rate": 1.8854840221160154e-06,
|
35428 |
+
"loss": 0.0455,
|
35429 |
+
"step": 174900
|
35430 |
+
},
|
35431 |
+
{
|
35432 |
+
"epoch": 49.18,
|
35433 |
+
"learning_rate": 1.8667416362102896e-06,
|
35434 |
+
"loss": 0.041,
|
35435 |
+
"step": 174930
|
35436 |
+
},
|
35437 |
+
{
|
35438 |
+
"epoch": 49.19,
|
35439 |
+
"learning_rate": 1.8479992503045638e-06,
|
35440 |
+
"loss": 0.0422,
|
35441 |
+
"step": 174960
|
35442 |
+
},
|
35443 |
+
{
|
35444 |
+
"epoch": 49.2,
|
35445 |
+
"learning_rate": 1.829256864398838e-06,
|
35446 |
+
"loss": 0.0407,
|
35447 |
+
"step": 174990
|
35448 |
+
},
|
35449 |
+
{
|
35450 |
+
"epoch": 49.2,
|
35451 |
+
"learning_rate": 1.8105144784931122e-06,
|
35452 |
+
"loss": 0.0414,
|
35453 |
+
"step": 175020
|
35454 |
+
},
|
35455 |
+
{
|
35456 |
+
"epoch": 49.21,
|
35457 |
+
"learning_rate": 1.7917720925873866e-06,
|
35458 |
+
"loss": 0.0444,
|
35459 |
+
"step": 175050
|
35460 |
+
},
|
35461 |
+
{
|
35462 |
+
"epoch": 49.22,
|
35463 |
+
"learning_rate": 1.7730297066816608e-06,
|
35464 |
+
"loss": 0.0374,
|
35465 |
+
"step": 175080
|
35466 |
+
},
|
35467 |
+
{
|
35468 |
+
"epoch": 49.23,
|
35469 |
+
"learning_rate": 1.7549120669727924e-06,
|
35470 |
+
"loss": 0.0429,
|
35471 |
+
"step": 175110
|
35472 |
+
},
|
35473 |
+
{
|
35474 |
+
"epoch": 49.24,
|
35475 |
+
"learning_rate": 1.7361696810670666e-06,
|
35476 |
+
"loss": 0.0431,
|
35477 |
+
"step": 175140
|
35478 |
+
},
|
35479 |
+
{
|
35480 |
+
"epoch": 49.25,
|
35481 |
+
"learning_rate": 1.7174272951613408e-06,
|
35482 |
+
"loss": 0.0397,
|
35483 |
+
"step": 175170
|
35484 |
+
},
|
35485 |
+
{
|
35486 |
+
"epoch": 49.25,
|
35487 |
+
"learning_rate": 1.698684909255615e-06,
|
35488 |
+
"loss": 0.0468,
|
35489 |
+
"step": 175200
|
35490 |
+
},
|
35491 |
+
{
|
35492 |
+
"epoch": 49.26,
|
35493 |
+
"learning_rate": 1.6799425233498892e-06,
|
35494 |
+
"loss": 0.0415,
|
35495 |
+
"step": 175230
|
35496 |
+
},
|
35497 |
+
{
|
35498 |
+
"epoch": 49.27,
|
35499 |
+
"learning_rate": 1.6612001374441632e-06,
|
35500 |
+
"loss": 0.0408,
|
35501 |
+
"step": 175260
|
35502 |
+
},
|
35503 |
+
{
|
35504 |
+
"epoch": 49.28,
|
35505 |
+
"learning_rate": 1.6424577515384378e-06,
|
35506 |
+
"loss": 0.0416,
|
35507 |
+
"step": 175290
|
35508 |
+
},
|
35509 |
+
{
|
35510 |
+
"epoch": 49.29,
|
35511 |
+
"learning_rate": 1.623715365632712e-06,
|
35512 |
+
"loss": 0.0403,
|
35513 |
+
"step": 175320
|
35514 |
+
},
|
35515 |
+
{
|
35516 |
+
"epoch": 49.3,
|
35517 |
+
"learning_rate": 1.6049729797269862e-06,
|
35518 |
+
"loss": 0.0445,
|
35519 |
+
"step": 175350
|
35520 |
+
},
|
35521 |
+
{
|
35522 |
+
"epoch": 49.31,
|
35523 |
+
"learning_rate": 1.5862305938212601e-06,
|
35524 |
+
"loss": 0.0398,
|
35525 |
+
"step": 175380
|
35526 |
+
},
|
35527 |
+
{
|
35528 |
+
"epoch": 49.31,
|
35529 |
+
"learning_rate": 1.5674882079155343e-06,
|
35530 |
+
"loss": 0.0413,
|
35531 |
+
"step": 175410
|
35532 |
+
},
|
35533 |
+
{
|
35534 |
+
"epoch": 49.32,
|
35535 |
+
"learning_rate": 1.5487458220098085e-06,
|
35536 |
+
"loss": 0.0442,
|
35537 |
+
"step": 175440
|
35538 |
+
},
|
35539 |
+
{
|
35540 |
+
"epoch": 49.33,
|
35541 |
+
"learning_rate": 1.530003436104083e-06,
|
35542 |
+
"loss": 0.0434,
|
35543 |
+
"step": 175470
|
35544 |
+
},
|
35545 |
+
{
|
35546 |
+
"epoch": 49.34,
|
35547 |
+
"learning_rate": 1.5112610501983571e-06,
|
35548 |
+
"loss": 0.0455,
|
35549 |
+
"step": 175500
|
35550 |
+
},
|
35551 |
+
{
|
35552 |
+
"epoch": 49.35,
|
35553 |
+
"learning_rate": 1.492518664292631e-06,
|
35554 |
+
"loss": 0.043,
|
35555 |
+
"step": 175530
|
35556 |
+
},
|
35557 |
+
{
|
35558 |
+
"epoch": 49.36,
|
35559 |
+
"learning_rate": 1.4737762783869053e-06,
|
35560 |
+
"loss": 0.0459,
|
35561 |
+
"step": 175560
|
35562 |
+
},
|
35563 |
+
{
|
35564 |
+
"epoch": 49.36,
|
35565 |
+
"learning_rate": 1.4550338924811797e-06,
|
35566 |
+
"loss": 0.044,
|
35567 |
+
"step": 175590
|
35568 |
+
},
|
35569 |
+
{
|
35570 |
+
"epoch": 49.37,
|
35571 |
+
"learning_rate": 1.4362915065754539e-06,
|
35572 |
+
"loss": 0.0394,
|
35573 |
+
"step": 175620
|
35574 |
+
},
|
35575 |
+
{
|
35576 |
+
"epoch": 49.38,
|
35577 |
+
"learning_rate": 1.417549120669728e-06,
|
35578 |
+
"loss": 0.0448,
|
35579 |
+
"step": 175650
|
35580 |
+
},
|
35581 |
+
{
|
35582 |
+
"epoch": 49.39,
|
35583 |
+
"learning_rate": 1.3988067347640023e-06,
|
35584 |
+
"loss": 0.0382,
|
35585 |
+
"step": 175680
|
35586 |
+
},
|
35587 |
+
{
|
35588 |
+
"epoch": 49.4,
|
35589 |
+
"learning_rate": 1.3800643488582764e-06,
|
35590 |
+
"loss": 0.0433,
|
35591 |
+
"step": 175710
|
35592 |
+
},
|
35593 |
+
{
|
35594 |
+
"epoch": 49.41,
|
35595 |
+
"learning_rate": 1.3613219629525506e-06,
|
35596 |
+
"loss": 0.0419,
|
35597 |
+
"step": 175740
|
35598 |
+
},
|
35599 |
+
{
|
35600 |
+
"epoch": 49.42,
|
35601 |
+
"learning_rate": 1.3425795770468248e-06,
|
35602 |
+
"loss": 0.039,
|
35603 |
+
"step": 175770
|
35604 |
+
},
|
35605 |
+
{
|
35606 |
+
"epoch": 49.42,
|
35607 |
+
"learning_rate": 1.323837191141099e-06,
|
35608 |
+
"loss": 0.0439,
|
35609 |
+
"step": 175800
|
35610 |
+
},
|
35611 |
+
{
|
35612 |
+
"epoch": 49.43,
|
35613 |
+
"learning_rate": 1.3050948052353732e-06,
|
35614 |
+
"loss": 0.0371,
|
35615 |
+
"step": 175830
|
35616 |
+
},
|
35617 |
+
{
|
35618 |
+
"epoch": 49.44,
|
35619 |
+
"learning_rate": 1.2863524193296474e-06,
|
35620 |
+
"loss": 0.0452,
|
35621 |
+
"step": 175860
|
35622 |
+
},
|
35623 |
+
{
|
35624 |
+
"epoch": 49.45,
|
35625 |
+
"learning_rate": 1.2676100334239216e-06,
|
35626 |
+
"loss": 0.0405,
|
35627 |
+
"step": 175890
|
35628 |
+
},
|
35629 |
+
{
|
35630 |
+
"epoch": 49.46,
|
35631 |
+
"learning_rate": 1.2488676475181958e-06,
|
35632 |
+
"loss": 0.0368,
|
35633 |
+
"step": 175920
|
35634 |
+
},
|
35635 |
+
{
|
35636 |
+
"epoch": 49.47,
|
35637 |
+
"learning_rate": 1.23012526161247e-06,
|
35638 |
+
"loss": 0.0468,
|
35639 |
+
"step": 175950
|
35640 |
+
},
|
35641 |
+
{
|
35642 |
+
"epoch": 49.47,
|
35643 |
+
"learning_rate": 1.2113828757067442e-06,
|
35644 |
+
"loss": 0.037,
|
35645 |
+
"step": 175980
|
35646 |
+
},
|
35647 |
+
{
|
35648 |
+
"epoch": 49.48,
|
35649 |
+
"learning_rate": 1.1926404898010184e-06,
|
35650 |
+
"loss": 0.0452,
|
35651 |
+
"step": 176010
|
35652 |
+
},
|
35653 |
+
{
|
35654 |
+
"epoch": 49.49,
|
35655 |
+
"learning_rate": 1.1738981038952925e-06,
|
35656 |
+
"loss": 0.0404,
|
35657 |
+
"step": 176040
|
35658 |
+
},
|
35659 |
+
{
|
35660 |
+
"epoch": 49.5,
|
35661 |
+
"learning_rate": 1.1551557179895667e-06,
|
35662 |
+
"loss": 0.0385,
|
35663 |
+
"step": 176070
|
35664 |
+
},
|
35665 |
+
{
|
35666 |
+
"epoch": 49.51,
|
35667 |
+
"learning_rate": 1.1364133320838411e-06,
|
35668 |
+
"loss": 0.0507,
|
35669 |
+
"step": 176100
|
35670 |
+
},
|
35671 |
+
{
|
35672 |
+
"epoch": 49.52,
|
35673 |
+
"learning_rate": 1.1176709461781151e-06,
|
35674 |
+
"loss": 0.0376,
|
35675 |
+
"step": 176130
|
35676 |
+
},
|
35677 |
+
{
|
35678 |
+
"epoch": 49.52,
|
35679 |
+
"learning_rate": 1.0989285602723893e-06,
|
35680 |
+
"loss": 0.0447,
|
35681 |
+
"step": 176160
|
35682 |
+
},
|
35683 |
+
{
|
35684 |
+
"epoch": 49.53,
|
35685 |
+
"learning_rate": 1.0801861743666635e-06,
|
35686 |
+
"loss": 0.0436,
|
35687 |
+
"step": 176190
|
35688 |
+
},
|
35689 |
+
{
|
35690 |
+
"epoch": 49.54,
|
35691 |
+
"learning_rate": 1.061443788460938e-06,
|
35692 |
+
"loss": 0.041,
|
35693 |
+
"step": 176220
|
35694 |
+
},
|
35695 |
+
{
|
35696 |
+
"epoch": 49.55,
|
35697 |
+
"learning_rate": 1.042701402555212e-06,
|
35698 |
+
"loss": 0.0448,
|
35699 |
+
"step": 176250
|
35700 |
+
},
|
35701 |
+
{
|
35702 |
+
"epoch": 49.56,
|
35703 |
+
"learning_rate": 1.023959016649486e-06,
|
35704 |
+
"loss": 0.0374,
|
35705 |
+
"step": 176280
|
35706 |
+
},
|
35707 |
+
{
|
35708 |
+
"epoch": 49.57,
|
35709 |
+
"learning_rate": 1.0052166307437603e-06,
|
35710 |
+
"loss": 0.0445,
|
35711 |
+
"step": 176310
|
35712 |
+
},
|
35713 |
+
{
|
35714 |
+
"epoch": 49.58,
|
35715 |
+
"learning_rate": 9.864742448380347e-07,
|
35716 |
+
"loss": 0.0399,
|
35717 |
+
"step": 176340
|
35718 |
+
},
|
35719 |
+
{
|
35720 |
+
"epoch": 49.58,
|
35721 |
+
"learning_rate": 9.677318589323089e-07,
|
35722 |
+
"loss": 0.0396,
|
35723 |
+
"step": 176370
|
35724 |
+
},
|
35725 |
+
{
|
35726 |
+
"epoch": 49.59,
|
35727 |
+
"learning_rate": 9.489894730265829e-07,
|
35728 |
+
"loss": 0.045,
|
35729 |
+
"step": 176400
|
35730 |
+
},
|
35731 |
+
{
|
35732 |
+
"epoch": 49.6,
|
35733 |
+
"learning_rate": 9.302470871208571e-07,
|
35734 |
+
"loss": 0.0401,
|
35735 |
+
"step": 176430
|
35736 |
+
},
|
35737 |
+
{
|
35738 |
+
"epoch": 49.61,
|
35739 |
+
"learning_rate": 9.115047012151314e-07,
|
35740 |
+
"loss": 0.0454,
|
35741 |
+
"step": 176460
|
35742 |
+
},
|
35743 |
+
{
|
35744 |
+
"epoch": 49.62,
|
35745 |
+
"learning_rate": 8.927623153094056e-07,
|
35746 |
+
"loss": 0.0412,
|
35747 |
+
"step": 176490
|
35748 |
+
},
|
35749 |
+
{
|
35750 |
+
"epoch": 49.63,
|
35751 |
+
"learning_rate": 8.740199294036798e-07,
|
35752 |
+
"loss": 0.0376,
|
35753 |
+
"step": 176520
|
35754 |
+
},
|
35755 |
+
{
|
35756 |
+
"epoch": 49.63,
|
35757 |
+
"learning_rate": 8.552775434979539e-07,
|
35758 |
+
"loss": 0.0451,
|
35759 |
+
"step": 176550
|
35760 |
+
},
|
35761 |
+
{
|
35762 |
+
"epoch": 49.64,
|
35763 |
+
"learning_rate": 8.365351575922283e-07,
|
35764 |
+
"loss": 0.0402,
|
35765 |
+
"step": 176580
|
35766 |
+
},
|
35767 |
+
{
|
35768 |
+
"epoch": 49.65,
|
35769 |
+
"learning_rate": 8.177927716865024e-07,
|
35770 |
+
"loss": 0.0433,
|
35771 |
+
"step": 176610
|
35772 |
+
},
|
35773 |
+
{
|
35774 |
+
"epoch": 49.66,
|
35775 |
+
"learning_rate": 7.990503857807766e-07,
|
35776 |
+
"loss": 0.0432,
|
35777 |
+
"step": 176640
|
35778 |
+
},
|
35779 |
+
{
|
35780 |
+
"epoch": 49.67,
|
35781 |
+
"learning_rate": 7.803079998750509e-07,
|
35782 |
+
"loss": 0.0424,
|
35783 |
+
"step": 176670
|
35784 |
+
},
|
35785 |
+
{
|
35786 |
+
"epoch": 49.68,
|
35787 |
+
"learning_rate": 7.61565613969325e-07,
|
35788 |
+
"loss": 0.0455,
|
35789 |
+
"step": 176700
|
35790 |
+
},
|
35791 |
+
{
|
35792 |
+
"epoch": 49.69,
|
35793 |
+
"learning_rate": 7.428232280635992e-07,
|
35794 |
+
"loss": 0.0404,
|
35795 |
+
"step": 176730
|
35796 |
+
},
|
35797 |
+
{
|
35798 |
+
"epoch": 49.69,
|
35799 |
+
"learning_rate": 7.240808421578733e-07,
|
35800 |
+
"loss": 0.0441,
|
35801 |
+
"step": 176760
|
35802 |
+
},
|
35803 |
+
{
|
35804 |
+
"epoch": 49.7,
|
35805 |
+
"learning_rate": 7.053384562521476e-07,
|
35806 |
+
"loss": 0.0422,
|
35807 |
+
"step": 176790
|
35808 |
+
},
|
35809 |
+
{
|
35810 |
+
"epoch": 49.71,
|
35811 |
+
"learning_rate": 6.865960703464218e-07,
|
35812 |
+
"loss": 0.0423,
|
35813 |
+
"step": 176820
|
35814 |
+
},
|
35815 |
+
{
|
35816 |
+
"epoch": 49.72,
|
35817 |
+
"learning_rate": 6.67853684440696e-07,
|
35818 |
+
"loss": 0.0415,
|
35819 |
+
"step": 176850
|
35820 |
+
},
|
35821 |
+
{
|
35822 |
+
"epoch": 49.73,
|
35823 |
+
"learning_rate": 6.491112985349702e-07,
|
35824 |
+
"loss": 0.0378,
|
35825 |
+
"step": 176880
|
35826 |
+
},
|
35827 |
+
{
|
35828 |
+
"epoch": 49.74,
|
35829 |
+
"learning_rate": 6.303689126292444e-07,
|
35830 |
+
"loss": 0.0427,
|
35831 |
+
"step": 176910
|
35832 |
+
},
|
35833 |
+
{
|
35834 |
+
"epoch": 49.74,
|
35835 |
+
"learning_rate": 6.116265267235186e-07,
|
35836 |
+
"loss": 0.0437,
|
35837 |
+
"step": 176940
|
35838 |
+
},
|
35839 |
+
{
|
35840 |
+
"epoch": 49.75,
|
35841 |
+
"learning_rate": 5.928841408177928e-07,
|
35842 |
+
"loss": 0.0383,
|
35843 |
+
"step": 176970
|
35844 |
+
},
|
35845 |
+
{
|
35846 |
+
"epoch": 49.76,
|
35847 |
+
"learning_rate": 5.74141754912067e-07,
|
35848 |
+
"loss": 0.0465,
|
35849 |
+
"step": 177000
|
35850 |
+
},
|
35851 |
+
{
|
35852 |
+
"epoch": 49.77,
|
35853 |
+
"learning_rate": 5.553993690063413e-07,
|
35854 |
+
"loss": 0.0377,
|
35855 |
+
"step": 177030
|
35856 |
+
},
|
35857 |
+
{
|
35858 |
+
"epoch": 49.78,
|
35859 |
+
"learning_rate": 5.366569831006153e-07,
|
35860 |
+
"loss": 0.0432,
|
35861 |
+
"step": 177060
|
35862 |
+
},
|
35863 |
+
{
|
35864 |
+
"epoch": 49.79,
|
35865 |
+
"learning_rate": 5.179145971948896e-07,
|
35866 |
+
"loss": 0.0445,
|
35867 |
+
"step": 177090
|
35868 |
+
},
|
35869 |
+
{
|
35870 |
+
"epoch": 49.79,
|
35871 |
+
"learning_rate": 4.991722112891638e-07,
|
35872 |
+
"loss": 0.0394,
|
35873 |
+
"step": 177120
|
35874 |
+
},
|
35875 |
+
{
|
35876 |
+
"epoch": 49.8,
|
35877 |
+
"learning_rate": 4.80429825383438e-07,
|
35878 |
+
"loss": 0.047,
|
35879 |
+
"step": 177150
|
35880 |
+
},
|
35881 |
+
{
|
35882 |
+
"epoch": 49.81,
|
35883 |
+
"learning_rate": 4.6168743947771216e-07,
|
35884 |
+
"loss": 0.0407,
|
35885 |
+
"step": 177180
|
35886 |
+
},
|
35887 |
+
{
|
35888 |
+
"epoch": 49.82,
|
35889 |
+
"learning_rate": 4.429450535719864e-07,
|
35890 |
+
"loss": 0.0439,
|
35891 |
+
"step": 177210
|
35892 |
+
},
|
35893 |
+
{
|
35894 |
+
"epoch": 49.83,
|
35895 |
+
"learning_rate": 4.242026676662606e-07,
|
35896 |
+
"loss": 0.0396,
|
35897 |
+
"step": 177240
|
35898 |
+
},
|
35899 |
+
{
|
35900 |
+
"epoch": 49.84,
|
35901 |
+
"learning_rate": 4.0546028176053484e-07,
|
35902 |
+
"loss": 0.0395,
|
35903 |
+
"step": 177270
|
35904 |
+
},
|
35905 |
+
{
|
35906 |
+
"epoch": 49.85,
|
35907 |
+
"learning_rate": 3.8671789585480897e-07,
|
35908 |
+
"loss": 0.0482,
|
35909 |
+
"step": 177300
|
35910 |
+
},
|
35911 |
+
{
|
35912 |
+
"epoch": 49.85,
|
35913 |
+
"learning_rate": 3.6797550994908316e-07,
|
35914 |
+
"loss": 0.0409,
|
35915 |
+
"step": 177330
|
35916 |
+
},
|
35917 |
+
{
|
35918 |
+
"epoch": 49.86,
|
35919 |
+
"learning_rate": 3.492331240433574e-07,
|
35920 |
+
"loss": 0.0444,
|
35921 |
+
"step": 177360
|
35922 |
+
},
|
35923 |
+
{
|
35924 |
+
"epoch": 49.87,
|
35925 |
+
"learning_rate": 3.304907381376316e-07,
|
35926 |
+
"loss": 0.0411,
|
35927 |
+
"step": 177390
|
35928 |
+
},
|
35929 |
+
{
|
35930 |
+
"epoch": 49.88,
|
35931 |
+
"learning_rate": 3.117483522319058e-07,
|
35932 |
+
"loss": 0.0408,
|
35933 |
+
"step": 177420
|
35934 |
+
},
|
35935 |
+
{
|
35936 |
+
"epoch": 49.89,
|
35937 |
+
"learning_rate": 2.9300596632618e-07,
|
35938 |
+
"loss": 0.0443,
|
35939 |
+
"step": 177450
|
35940 |
+
},
|
35941 |
+
{
|
35942 |
+
"epoch": 49.9,
|
35943 |
+
"learning_rate": 2.7426358042045417e-07,
|
35944 |
+
"loss": 0.0386,
|
35945 |
+
"step": 177480
|
35946 |
+
},
|
35947 |
+
{
|
35948 |
+
"epoch": 49.9,
|
35949 |
+
"learning_rate": 2.5552119451472836e-07,
|
35950 |
+
"loss": 0.0462,
|
35951 |
+
"step": 177510
|
35952 |
+
},
|
35953 |
+
{
|
35954 |
+
"epoch": 49.91,
|
35955 |
+
"learning_rate": 2.3677880860900258e-07,
|
35956 |
+
"loss": 0.042,
|
35957 |
+
"step": 177540
|
35958 |
+
},
|
35959 |
+
{
|
35960 |
+
"epoch": 49.92,
|
35961 |
+
"learning_rate": 2.180364227032768e-07,
|
35962 |
+
"loss": 0.0422,
|
35963 |
+
"step": 177570
|
35964 |
+
},
|
35965 |
+
{
|
35966 |
+
"epoch": 49.93,
|
35967 |
+
"learning_rate": 1.9929403679755099e-07,
|
35968 |
+
"loss": 0.043,
|
35969 |
+
"step": 177600
|
35970 |
+
},
|
35971 |
+
{
|
35972 |
+
"epoch": 49.94,
|
35973 |
+
"learning_rate": 1.805516508918252e-07,
|
35974 |
+
"loss": 0.0374,
|
35975 |
+
"step": 177630
|
35976 |
+
},
|
35977 |
+
{
|
35978 |
+
"epoch": 49.95,
|
35979 |
+
"learning_rate": 1.6180926498609942e-07,
|
35980 |
+
"loss": 0.043,
|
35981 |
+
"step": 177660
|
35982 |
+
},
|
35983 |
+
{
|
35984 |
+
"epoch": 49.95,
|
35985 |
+
"learning_rate": 1.430668790803736e-07,
|
35986 |
+
"loss": 0.0419,
|
35987 |
+
"step": 177690
|
35988 |
+
},
|
35989 |
+
{
|
35990 |
+
"epoch": 49.96,
|
35991 |
+
"learning_rate": 1.243244931746478e-07,
|
35992 |
+
"loss": 0.0397,
|
35993 |
+
"step": 177720
|
35994 |
+
},
|
35995 |
+
{
|
35996 |
+
"epoch": 49.97,
|
35997 |
+
"learning_rate": 1.05582107268922e-07,
|
35998 |
+
"loss": 0.0488,
|
35999 |
+
"step": 177750
|
36000 |
+
},
|
36001 |
+
{
|
36002 |
+
"epoch": 49.98,
|
36003 |
+
"learning_rate": 8.68397213631962e-08,
|
36004 |
+
"loss": 0.0399,
|
36005 |
+
"step": 177780
|
36006 |
+
},
|
36007 |
+
{
|
36008 |
+
"epoch": 49.99,
|
36009 |
+
"learning_rate": 6.80973354574704e-08,
|
36010 |
+
"loss": 0.0426,
|
36011 |
+
"step": 177810
|
36012 |
+
},
|
36013 |
+
{
|
36014 |
+
"epoch": 50.0,
|
36015 |
+
"learning_rate": 4.9354949551744604e-08,
|
36016 |
+
"loss": 0.0428,
|
36017 |
+
"step": 177840
|
36018 |
+
},
|
36019 |
+
{
|
36020 |
+
"epoch": 50.0,
|
36021 |
+
"eval_loss": 0.11027602851390839,
|
36022 |
+
"eval_runtime": 645.5008,
|
36023 |
+
"eval_samples_per_second": 24.838,
|
36024 |
+
"eval_steps_per_second": 0.778,
|
36025 |
+
"eval_wer": 0.0915207511634572,
|
36026 |
+
"step": 177850
|
36027 |
+
},
|
36028 |
+
{
|
36029 |
+
"epoch": 50.0,
|
36030 |
+
"step": 177850,
|
36031 |
+
"total_flos": 3.938373027563576e+21,
|
36032 |
+
"train_loss": 0.1478050081201038,
|
36033 |
+
"train_runtime": 1439234.2576,
|
36034 |
+
"train_samples_per_second": 15.819,
|
36035 |
+
"train_steps_per_second": 0.124
|
36036 |
}
|
36037 |
],
|
36038 |
"max_steps": 177850,
|
36039 |
"num_train_epochs": 50,
|
36040 |
+
"total_flos": 3.938373027563576e+21,
|
36041 |
"trial_name": null,
|
36042 |
"trial_params": null
|
36043 |
}
|