jimjakdiend commited on
Commit
3cb5768
1 Parent(s): 5c4e58a

End of training

Browse files
Files changed (3) hide show
  1. README.md +58 -0
  2. generation_config.json +197 -0
  3. model.safetensors +1 -1
README.md ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ base_model: distil-whisper/distil-large-v3
4
+ tags:
5
+ - generated_from_trainer
6
+ model-index:
7
+ - name: distilwhisper_finetune
8
+ results: []
9
+ ---
10
+
11
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
+ should probably proofread and complete it, then remove this comment. -->
13
+
14
+ # distilwhisper_finetune
15
+
16
+ This model is a fine-tuned version of [distil-whisper/distil-large-v3](https://huggingface.co/distil-whisper/distil-large-v3) on an unknown dataset.
17
+ It achieves the following results on the evaluation set:
18
+ - eval_loss: 0.7639
19
+ - eval_wer: 30.1396
20
+ - eval_runtime: 848.6992
21
+ - eval_samples_per_second: 0.825
22
+ - eval_steps_per_second: 0.104
23
+ - epoch: 0.1143
24
+ - step: 20
25
+
26
+ ## Model description
27
+
28
+ More information needed
29
+
30
+ ## Intended uses & limitations
31
+
32
+ More information needed
33
+
34
+ ## Training and evaluation data
35
+
36
+ More information needed
37
+
38
+ ## Training procedure
39
+
40
+ ### Training hyperparameters
41
+
42
+ The following hyperparameters were used during training:
43
+ - learning_rate: 1e-05
44
+ - train_batch_size: 16
45
+ - eval_batch_size: 8
46
+ - seed: 42
47
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
48
+ - lr_scheduler_type: linear
49
+ - lr_scheduler_warmup_steps: 500
50
+ - training_steps: 5000
51
+ - mixed_precision_training: Native AMP
52
+
53
+ ### Framework versions
54
+
55
+ - Transformers 4.41.0
56
+ - Pytorch 2.2.1+cu121
57
+ - Datasets 2.19.1
58
+ - Tokenizers 0.19.1
generation_config.json ADDED
@@ -0,0 +1,197 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alignment_heads": [
3
+ [
4
+ 1,
5
+ 0
6
+ ],
7
+ [
8
+ 1,
9
+ 1
10
+ ],
11
+ [
12
+ 1,
13
+ 2
14
+ ],
15
+ [
16
+ 1,
17
+ 3
18
+ ],
19
+ [
20
+ 1,
21
+ 4
22
+ ],
23
+ [
24
+ 1,
25
+ 5
26
+ ],
27
+ [
28
+ 1,
29
+ 6
30
+ ],
31
+ [
32
+ 1,
33
+ 7
34
+ ],
35
+ [
36
+ 1,
37
+ 8
38
+ ],
39
+ [
40
+ 1,
41
+ 9
42
+ ],
43
+ [
44
+ 1,
45
+ 10
46
+ ],
47
+ [
48
+ 1,
49
+ 11
50
+ ],
51
+ [
52
+ 1,
53
+ 12
54
+ ],
55
+ [
56
+ 1,
57
+ 13
58
+ ],
59
+ [
60
+ 1,
61
+ 14
62
+ ],
63
+ [
64
+ 1,
65
+ 15
66
+ ],
67
+ [
68
+ 1,
69
+ 16
70
+ ],
71
+ [
72
+ 1,
73
+ 17
74
+ ],
75
+ [
76
+ 1,
77
+ 18
78
+ ],
79
+ [
80
+ 1,
81
+ 19
82
+ ]
83
+ ],
84
+ "begin_suppress_tokens": [
85
+ 220,
86
+ 50257
87
+ ],
88
+ "bos_token_id": 50257,
89
+ "decoder_start_token_id": 50258,
90
+ "eos_token_id": 50257,
91
+ "is_multilingual": true,
92
+ "lang_to_id": {
93
+ "<|en|>": 50259
94
+ },
95
+ "language": "english",
96
+ "max_initial_timestamp_index": 50,
97
+ "max_length": 448,
98
+ "no_timestamps_token_id": 50364,
99
+ "pad_token_id": 50257,
100
+ "prev_sot_token_id": 50362,
101
+ "return_timestamps": false,
102
+ "suppress_tokens": [
103
+ 1,
104
+ 2,
105
+ 7,
106
+ 8,
107
+ 9,
108
+ 10,
109
+ 14,
110
+ 25,
111
+ 26,
112
+ 27,
113
+ 28,
114
+ 29,
115
+ 31,
116
+ 58,
117
+ 59,
118
+ 60,
119
+ 61,
120
+ 62,
121
+ 63,
122
+ 90,
123
+ 91,
124
+ 92,
125
+ 93,
126
+ 359,
127
+ 503,
128
+ 522,
129
+ 542,
130
+ 873,
131
+ 893,
132
+ 902,
133
+ 918,
134
+ 922,
135
+ 931,
136
+ 1350,
137
+ 1853,
138
+ 1982,
139
+ 2460,
140
+ 2627,
141
+ 3246,
142
+ 3253,
143
+ 3268,
144
+ 3536,
145
+ 3846,
146
+ 3961,
147
+ 4183,
148
+ 4667,
149
+ 6585,
150
+ 6647,
151
+ 7273,
152
+ 9061,
153
+ 9383,
154
+ 10428,
155
+ 10929,
156
+ 11938,
157
+ 12033,
158
+ 12331,
159
+ 12562,
160
+ 13793,
161
+ 14157,
162
+ 14635,
163
+ 15265,
164
+ 15618,
165
+ 16553,
166
+ 16604,
167
+ 18362,
168
+ 18956,
169
+ 20075,
170
+ 21675,
171
+ 22520,
172
+ 26130,
173
+ 26161,
174
+ 26435,
175
+ 28279,
176
+ 29464,
177
+ 31650,
178
+ 32302,
179
+ 32470,
180
+ 36865,
181
+ 42863,
182
+ 47425,
183
+ 49870,
184
+ 50254,
185
+ 50258,
186
+ 50359,
187
+ 50360,
188
+ 50361,
189
+ 50362,
190
+ 50363
191
+ ],
192
+ "task": "transcribe",
193
+ "task_to_id": {
194
+ "transcribe": 50359
195
+ },
196
+ "transformers_version": "4.41.0"
197
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bcb6690ce64294a67ae991f8c53b688ee0c3e850a762c025b63749ae8a52afc0
3
  size 3025686376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:316c91d4a31204de325f265c4e68315ac453f7b3e11b903e26f47ac954ad61a7
3
  size 3025686376