pravin96 commited on
Commit
26f015e
1 Parent(s): 28c6841

Model save

Browse files
Files changed (2) hide show
  1. README.md +5 -5
  2. generation_config.json +74 -271
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: transformers
3
  license: mit
4
- base_model: distil-whisper/distil-large-v3
5
  tags:
6
  - generated_from_trainer
7
  datasets:
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # distil_whisper_en
18
 
19
- This model is a fine-tuned version of [distil-whisper/distil-large-v3](https://huggingface.co/distil-whisper/distil-large-v3) on the generator dataset.
20
 
21
  ## Model description
22
 
@@ -36,15 +36,15 @@ More information needed
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 0.0001
39
- - train_batch_size: 4
40
  - eval_batch_size: 8
41
  - seed: 42
42
  - gradient_accumulation_steps: 2
43
- - total_train_batch_size: 8
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - lr_scheduler_warmup_steps: 500
47
- - training_steps: 200
48
  - mixed_precision_training: Native AMP
49
 
50
  ### Training results
 
1
  ---
2
  library_name: transformers
3
  license: mit
4
+ base_model: distil-whisper/distil-medium.en
5
  tags:
6
  - generated_from_trainer
7
  datasets:
 
16
 
17
  # distil_whisper_en
18
 
19
+ This model is a fine-tuned version of [distil-whisper/distil-medium.en](https://huggingface.co/distil-whisper/distil-medium.en) on the generator dataset.
20
 
21
  ## Model description
22
 
 
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 0.0001
39
+ - train_batch_size: 2
40
  - eval_batch_size: 8
41
  - seed: 42
42
  - gradient_accumulation_steps: 2
43
+ - total_train_batch_size: 4
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - lr_scheduler_warmup_steps: 500
47
+ - training_steps: 100
48
  - mixed_precision_training: Native AMP
49
 
50
  ### Training results
generation_config.json CHANGED
@@ -1,212 +1,17 @@
1
  {
2
- "alignment_heads": [
3
- [
4
- 1,
5
- 0
6
- ],
7
- [
8
- 1,
9
- 1
10
- ],
11
- [
12
- 1,
13
- 2
14
- ],
15
- [
16
- 1,
17
- 3
18
- ],
19
- [
20
- 1,
21
- 4
22
- ],
23
- [
24
- 1,
25
- 5
26
- ],
27
- [
28
- 1,
29
- 6
30
- ],
31
- [
32
- 1,
33
- 7
34
- ],
35
- [
36
- 1,
37
- 8
38
- ],
39
- [
40
- 1,
41
- 9
42
- ],
43
- [
44
- 1,
45
- 10
46
- ],
47
- [
48
- 1,
49
- 11
50
- ],
51
- [
52
- 1,
53
- 12
54
- ],
55
- [
56
- 1,
57
- 13
58
- ],
59
- [
60
- 1,
61
- 14
62
- ],
63
- [
64
- 1,
65
- 15
66
- ],
67
- [
68
- 1,
69
- 16
70
- ],
71
- [
72
- 1,
73
- 17
74
- ],
75
- [
76
- 1,
77
- 18
78
- ],
79
- [
80
- 1,
81
- 19
82
- ]
83
- ],
84
  "begin_suppress_tokens": [
85
  220,
86
- 50257
87
  ],
88
  "bos_token_id": 50257,
89
- "decoder_start_token_id": 50258,
90
- "eos_token_id": 50257,
91
- "forced_decoder_ids": [
92
- [
93
- 1,
94
- null
95
- ],
96
- [
97
- 2,
98
- 50360
99
- ]
100
- ],
101
- "is_multilingual": true,
102
- "lang_to_id": {
103
- "<|af|>": 50327,
104
- "<|am|>": 50334,
105
- "<|ar|>": 50272,
106
- "<|as|>": 50350,
107
- "<|az|>": 50304,
108
- "<|ba|>": 50355,
109
- "<|be|>": 50330,
110
- "<|bg|>": 50292,
111
- "<|bn|>": 50302,
112
- "<|bo|>": 50347,
113
- "<|br|>": 50309,
114
- "<|bs|>": 50315,
115
- "<|ca|>": 50270,
116
- "<|cs|>": 50283,
117
- "<|cy|>": 50297,
118
- "<|da|>": 50285,
119
- "<|de|>": 50261,
120
- "<|el|>": 50281,
121
- "<|en|>": 50259,
122
- "<|es|>": 50262,
123
- "<|et|>": 50307,
124
- "<|eu|>": 50310,
125
- "<|fa|>": 50300,
126
- "<|fi|>": 50277,
127
- "<|fo|>": 50338,
128
- "<|fr|>": 50265,
129
- "<|gl|>": 50319,
130
- "<|gu|>": 50333,
131
- "<|haw|>": 50352,
132
- "<|ha|>": 50354,
133
- "<|he|>": 50279,
134
- "<|hi|>": 50276,
135
- "<|hr|>": 50291,
136
- "<|ht|>": 50339,
137
- "<|hu|>": 50286,
138
- "<|hy|>": 50312,
139
- "<|id|>": 50275,
140
- "<|is|>": 50311,
141
- "<|it|>": 50274,
142
- "<|ja|>": 50266,
143
- "<|jw|>": 50356,
144
- "<|ka|>": 50329,
145
- "<|kk|>": 50316,
146
- "<|km|>": 50323,
147
- "<|kn|>": 50306,
148
- "<|ko|>": 50264,
149
- "<|la|>": 50294,
150
- "<|lb|>": 50345,
151
- "<|ln|>": 50353,
152
- "<|lo|>": 50336,
153
- "<|lt|>": 50293,
154
- "<|lv|>": 50301,
155
- "<|mg|>": 50349,
156
- "<|mi|>": 50295,
157
- "<|mk|>": 50308,
158
- "<|ml|>": 50296,
159
- "<|mn|>": 50314,
160
- "<|mr|>": 50320,
161
- "<|ms|>": 50282,
162
- "<|mt|>": 50343,
163
- "<|my|>": 50346,
164
- "<|ne|>": 50313,
165
- "<|nl|>": 50271,
166
- "<|nn|>": 50342,
167
- "<|no|>": 50288,
168
- "<|oc|>": 50328,
169
- "<|pa|>": 50321,
170
- "<|pl|>": 50269,
171
- "<|ps|>": 50340,
172
- "<|pt|>": 50267,
173
- "<|ro|>": 50284,
174
- "<|ru|>": 50263,
175
- "<|sa|>": 50344,
176
- "<|sd|>": 50332,
177
- "<|si|>": 50322,
178
- "<|sk|>": 50298,
179
- "<|sl|>": 50305,
180
- "<|sn|>": 50324,
181
- "<|so|>": 50326,
182
- "<|sq|>": 50317,
183
- "<|sr|>": 50303,
184
- "<|su|>": 50357,
185
- "<|sv|>": 50273,
186
- "<|sw|>": 50318,
187
- "<|ta|>": 50287,
188
- "<|te|>": 50299,
189
- "<|tg|>": 50331,
190
- "<|th|>": 50289,
191
- "<|tk|>": 50341,
192
- "<|tl|>": 50348,
193
- "<|tr|>": 50268,
194
- "<|tt|>": 50351,
195
- "<|uk|>": 50280,
196
- "<|ur|>": 50290,
197
- "<|uz|>": 50337,
198
- "<|vi|>": 50278,
199
- "<|yi|>": 50335,
200
- "<|yo|>": 50325,
201
- "<|yue|>": 50358,
202
- "<|zh|>": 50260
203
- },
204
- "language": "<|en|>",
205
  "max_initial_timestamp_index": 50,
206
  "max_length": 448,
207
- "no_timestamps_token_id": 50364,
208
- "pad_token_id": 50257,
209
- "prev_sot_token_id": 50362,
210
  "return_timestamps": false,
211
  "suppress_tokens": [
212
  1,
@@ -232,76 +37,74 @@
232
  91,
233
  92,
234
  93,
235
- 359,
236
- 503,
237
- 522,
238
- 542,
239
- 873,
240
- 893,
241
- 902,
242
- 918,
243
- 922,
244
- 931,
245
- 1350,
246
- 1853,
247
- 1982,
248
- 2460,
249
- 2627,
250
- 3246,
251
- 3253,
252
- 3268,
253
- 3536,
254
- 3846,
255
- 3961,
256
- 4183,
257
- 4667,
258
- 6585,
259
- 6647,
260
- 7273,
261
- 9061,
262
- 9383,
263
- 10428,
264
- 10929,
265
- 11938,
266
- 12033,
267
- 12331,
268
- 12562,
269
- 13793,
270
- 14157,
271
- 14635,
272
- 15265,
273
- 15618,
274
- 16553,
275
- 16604,
276
- 18362,
277
- 18956,
278
- 20075,
279
- 21675,
280
- 22520,
281
- 26130,
282
- 26161,
283
- 26435,
284
- 28279,
285
- 29464,
286
- 31650,
287
- 32302,
288
- 32470,
289
- 36865,
290
- 42863,
291
- 47425,
292
- 49870,
293
- 50254,
294
- 50258,
 
 
 
 
295
  50359,
296
  50360,
297
- 50361,
298
- 50362,
299
- 50363
300
  ],
301
- "task": "transcribe",
302
- "task_to_id": {
303
- "transcribe": 50360,
304
- "translate": 50359
305
- },
306
- "transformers_version": "4.44.2"
307
  }
 
1
  {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  "begin_suppress_tokens": [
3
  220,
4
+ 50256
5
  ],
6
  "bos_token_id": 50257,
7
+ "decoder_start_token_id": 50257,
8
+ "eos_token_id": 50256,
9
+ "is_multilingual": false,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  "max_initial_timestamp_index": 50,
11
  "max_length": 448,
12
+ "no_timestamps_token_id": 50362,
13
+ "pad_token_id": 50256,
14
+ "prev_sot_token_id": 50360,
15
  "return_timestamps": false,
16
  "suppress_tokens": [
17
  1,
 
37
  91,
38
  92,
39
  93,
40
+ 357,
41
+ 366,
42
+ 438,
43
+ 532,
44
+ 685,
45
+ 705,
46
+ 796,
47
+ 930,
48
+ 1058,
49
+ 1220,
50
+ 1267,
51
+ 1279,
52
+ 1303,
53
+ 1343,
54
+ 1377,
55
+ 1391,
56
+ 1635,
57
+ 1782,
58
+ 1875,
59
+ 2162,
60
+ 2361,
61
+ 2488,
62
+ 3467,
63
+ 4008,
64
+ 4211,
65
+ 4600,
66
+ 4808,
67
+ 5299,
68
+ 5855,
69
+ 6329,
70
+ 7203,
71
+ 9609,
72
+ 9959,
73
+ 10563,
74
+ 10786,
75
+ 11420,
76
+ 11709,
77
+ 11907,
78
+ 13163,
79
+ 13697,
80
+ 13700,
81
+ 14808,
82
+ 15306,
83
+ 16410,
84
+ 16791,
85
+ 17992,
86
+ 19203,
87
+ 19510,
88
+ 20724,
89
+ 22305,
90
+ 22935,
91
+ 27007,
92
+ 30109,
93
+ 30420,
94
+ 33409,
95
+ 34949,
96
+ 40283,
97
+ 40493,
98
+ 40549,
99
+ 47282,
100
+ 49146,
101
+ 50257,
102
+ 50357,
103
+ 50358,
104
  50359,
105
  50360,
106
+ 50361
 
 
107
  ],
108
+ "transformers_version": "4.44.2",
109
+ "use_scan": false
 
 
 
 
110
  }