aiola commited on
Commit
a57e12f
1 Parent(s): 5fe2021

Upload model.safetensors.index.json

Browse files
Files changed (1) hide show
  1. model.safetensors.index.json +3 -35
model.safetensors.index.json CHANGED
@@ -1,12 +1,12 @@
1
  {
2
  "metadata": {
3
- "total_size": 6173383680
4
  },
5
  "weight_map": {
6
  "model.decoder.embed_positions.weight": "model-00001-of-00002.safetensors",
7
  "model.decoder.embed_tokens.weight": "model-00001-of-00002.safetensors",
8
- "model.decoder.layer_norm.bias": "model-00001-of-00002.safetensors",
9
- "model.decoder.layer_norm.weight": "model-00001-of-00002.safetensors",
10
  "model.decoder.layers.0.encoder_attn.k_proj.weight": "model-00001-of-00002.safetensors",
11
  "model.decoder.layers.0.encoder_attn.out_proj.bias": "model-00001-of-00002.safetensors",
12
  "model.decoder.layers.0.encoder_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -22,7 +22,6 @@
22
  "model.decoder.layers.0.fc2.weight": "model-00001-of-00002.safetensors",
23
  "model.decoder.layers.0.final_layer_norm.bias": "model-00001-of-00002.safetensors",
24
  "model.decoder.layers.0.final_layer_norm.weight": "model-00001-of-00002.safetensors",
25
- "model.decoder.layers.0.no_speech_bias": "model-00001-of-00002.safetensors",
26
  "model.decoder.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
27
  "model.decoder.layers.0.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
28
  "model.decoder.layers.0.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -47,7 +46,6 @@
47
  "model.decoder.layers.1.fc2.weight": "model-00001-of-00002.safetensors",
48
  "model.decoder.layers.1.final_layer_norm.bias": "model-00001-of-00002.safetensors",
49
  "model.decoder.layers.1.final_layer_norm.weight": "model-00001-of-00002.safetensors",
50
- "model.decoder.layers.1.no_speech_bias": "model-00001-of-00002.safetensors",
51
  "model.decoder.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
52
  "model.decoder.layers.1.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
53
  "model.decoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -72,7 +70,6 @@
72
  "model.decoder.layers.10.fc2.weight": "model-00001-of-00002.safetensors",
73
  "model.decoder.layers.10.final_layer_norm.bias": "model-00001-of-00002.safetensors",
74
  "model.decoder.layers.10.final_layer_norm.weight": "model-00001-of-00002.safetensors",
75
- "model.decoder.layers.10.no_speech_bias": "model-00001-of-00002.safetensors",
76
  "model.decoder.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
77
  "model.decoder.layers.10.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
78
  "model.decoder.layers.10.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -97,7 +94,6 @@
97
  "model.decoder.layers.11.fc2.weight": "model-00001-of-00002.safetensors",
98
  "model.decoder.layers.11.final_layer_norm.bias": "model-00001-of-00002.safetensors",
99
  "model.decoder.layers.11.final_layer_norm.weight": "model-00001-of-00002.safetensors",
100
- "model.decoder.layers.11.no_speech_bias": "model-00001-of-00002.safetensors",
101
  "model.decoder.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
102
  "model.decoder.layers.11.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
103
  "model.decoder.layers.11.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -122,7 +118,6 @@
122
  "model.decoder.layers.12.fc2.weight": "model-00001-of-00002.safetensors",
123
  "model.decoder.layers.12.final_layer_norm.bias": "model-00001-of-00002.safetensors",
124
  "model.decoder.layers.12.final_layer_norm.weight": "model-00001-of-00002.safetensors",
125
- "model.decoder.layers.12.no_speech_bias": "model-00001-of-00002.safetensors",
126
  "model.decoder.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
127
  "model.decoder.layers.12.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
128
  "model.decoder.layers.12.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -147,7 +142,6 @@
147
  "model.decoder.layers.13.fc2.weight": "model-00001-of-00002.safetensors",
148
  "model.decoder.layers.13.final_layer_norm.bias": "model-00001-of-00002.safetensors",
149
  "model.decoder.layers.13.final_layer_norm.weight": "model-00001-of-00002.safetensors",
150
- "model.decoder.layers.13.no_speech_bias": "model-00001-of-00002.safetensors",
151
  "model.decoder.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
152
  "model.decoder.layers.13.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
153
  "model.decoder.layers.13.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -172,7 +166,6 @@
172
  "model.decoder.layers.14.fc2.weight": "model-00001-of-00002.safetensors",
173
  "model.decoder.layers.14.final_layer_norm.bias": "model-00001-of-00002.safetensors",
174
  "model.decoder.layers.14.final_layer_norm.weight": "model-00001-of-00002.safetensors",
175
- "model.decoder.layers.14.no_speech_bias": "model-00001-of-00002.safetensors",
176
  "model.decoder.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
177
  "model.decoder.layers.14.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
178
  "model.decoder.layers.14.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -197,7 +190,6 @@
197
  "model.decoder.layers.15.fc2.weight": "model-00001-of-00002.safetensors",
198
  "model.decoder.layers.15.final_layer_norm.bias": "model-00001-of-00002.safetensors",
199
  "model.decoder.layers.15.final_layer_norm.weight": "model-00001-of-00002.safetensors",
200
- "model.decoder.layers.15.no_speech_bias": "model-00001-of-00002.safetensors",
201
  "model.decoder.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
202
  "model.decoder.layers.15.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
203
  "model.decoder.layers.15.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -222,7 +214,6 @@
222
  "model.decoder.layers.16.fc2.weight": "model-00001-of-00002.safetensors",
223
  "model.decoder.layers.16.final_layer_norm.bias": "model-00001-of-00002.safetensors",
224
  "model.decoder.layers.16.final_layer_norm.weight": "model-00001-of-00002.safetensors",
225
- "model.decoder.layers.16.no_speech_bias": "model-00001-of-00002.safetensors",
226
  "model.decoder.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
227
  "model.decoder.layers.16.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
228
  "model.decoder.layers.16.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -247,7 +238,6 @@
247
  "model.decoder.layers.17.fc2.weight": "model-00001-of-00002.safetensors",
248
  "model.decoder.layers.17.final_layer_norm.bias": "model-00001-of-00002.safetensors",
249
  "model.decoder.layers.17.final_layer_norm.weight": "model-00001-of-00002.safetensors",
250
- "model.decoder.layers.17.no_speech_bias": "model-00001-of-00002.safetensors",
251
  "model.decoder.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
252
  "model.decoder.layers.17.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
253
  "model.decoder.layers.17.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -272,7 +262,6 @@
272
  "model.decoder.layers.18.fc2.weight": "model-00001-of-00002.safetensors",
273
  "model.decoder.layers.18.final_layer_norm.bias": "model-00001-of-00002.safetensors",
274
  "model.decoder.layers.18.final_layer_norm.weight": "model-00001-of-00002.safetensors",
275
- "model.decoder.layers.18.no_speech_bias": "model-00001-of-00002.safetensors",
276
  "model.decoder.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
277
  "model.decoder.layers.18.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
278
  "model.decoder.layers.18.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -297,7 +286,6 @@
297
  "model.decoder.layers.19.fc2.weight": "model-00001-of-00002.safetensors",
298
  "model.decoder.layers.19.final_layer_norm.bias": "model-00001-of-00002.safetensors",
299
  "model.decoder.layers.19.final_layer_norm.weight": "model-00001-of-00002.safetensors",
300
- "model.decoder.layers.19.no_speech_bias": "model-00001-of-00002.safetensors",
301
  "model.decoder.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
302
  "model.decoder.layers.19.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
303
  "model.decoder.layers.19.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -322,7 +310,6 @@
322
  "model.decoder.layers.2.fc2.weight": "model-00001-of-00002.safetensors",
323
  "model.decoder.layers.2.final_layer_norm.bias": "model-00001-of-00002.safetensors",
324
  "model.decoder.layers.2.final_layer_norm.weight": "model-00001-of-00002.safetensors",
325
- "model.decoder.layers.2.no_speech_bias": "model-00001-of-00002.safetensors",
326
  "model.decoder.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
327
  "model.decoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
328
  "model.decoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -347,7 +334,6 @@
347
  "model.decoder.layers.20.fc2.weight": "model-00002-of-00002.safetensors",
348
  "model.decoder.layers.20.final_layer_norm.bias": "model-00002-of-00002.safetensors",
349
  "model.decoder.layers.20.final_layer_norm.weight": "model-00002-of-00002.safetensors",
350
- "model.decoder.layers.20.no_speech_bias": "model-00001-of-00002.safetensors",
351
  "model.decoder.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
352
  "model.decoder.layers.20.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
353
  "model.decoder.layers.20.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -372,7 +358,6 @@
372
  "model.decoder.layers.21.fc2.weight": "model-00002-of-00002.safetensors",
373
  "model.decoder.layers.21.final_layer_norm.bias": "model-00002-of-00002.safetensors",
374
  "model.decoder.layers.21.final_layer_norm.weight": "model-00002-of-00002.safetensors",
375
- "model.decoder.layers.21.no_speech_bias": "model-00002-of-00002.safetensors",
376
  "model.decoder.layers.21.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
377
  "model.decoder.layers.21.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
378
  "model.decoder.layers.21.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
@@ -397,7 +382,6 @@
397
  "model.decoder.layers.22.fc2.weight": "model-00002-of-00002.safetensors",
398
  "model.decoder.layers.22.final_layer_norm.bias": "model-00002-of-00002.safetensors",
399
  "model.decoder.layers.22.final_layer_norm.weight": "model-00002-of-00002.safetensors",
400
- "model.decoder.layers.22.no_speech_bias": "model-00002-of-00002.safetensors",
401
  "model.decoder.layers.22.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
402
  "model.decoder.layers.22.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
403
  "model.decoder.layers.22.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
@@ -422,7 +406,6 @@
422
  "model.decoder.layers.23.fc2.weight": "model-00002-of-00002.safetensors",
423
  "model.decoder.layers.23.final_layer_norm.bias": "model-00002-of-00002.safetensors",
424
  "model.decoder.layers.23.final_layer_norm.weight": "model-00002-of-00002.safetensors",
425
- "model.decoder.layers.23.no_speech_bias": "model-00002-of-00002.safetensors",
426
  "model.decoder.layers.23.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
427
  "model.decoder.layers.23.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
428
  "model.decoder.layers.23.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
@@ -447,7 +430,6 @@
447
  "model.decoder.layers.24.fc2.weight": "model-00002-of-00002.safetensors",
448
  "model.decoder.layers.24.final_layer_norm.bias": "model-00002-of-00002.safetensors",
449
  "model.decoder.layers.24.final_layer_norm.weight": "model-00002-of-00002.safetensors",
450
- "model.decoder.layers.24.no_speech_bias": "model-00002-of-00002.safetensors",
451
  "model.decoder.layers.24.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
452
  "model.decoder.layers.24.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
453
  "model.decoder.layers.24.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
@@ -472,7 +454,6 @@
472
  "model.decoder.layers.25.fc2.weight": "model-00002-of-00002.safetensors",
473
  "model.decoder.layers.25.final_layer_norm.bias": "model-00002-of-00002.safetensors",
474
  "model.decoder.layers.25.final_layer_norm.weight": "model-00002-of-00002.safetensors",
475
- "model.decoder.layers.25.no_speech_bias": "model-00002-of-00002.safetensors",
476
  "model.decoder.layers.25.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
477
  "model.decoder.layers.25.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
478
  "model.decoder.layers.25.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
@@ -497,7 +478,6 @@
497
  "model.decoder.layers.26.fc2.weight": "model-00002-of-00002.safetensors",
498
  "model.decoder.layers.26.final_layer_norm.bias": "model-00002-of-00002.safetensors",
499
  "model.decoder.layers.26.final_layer_norm.weight": "model-00002-of-00002.safetensors",
500
- "model.decoder.layers.26.no_speech_bias": "model-00002-of-00002.safetensors",
501
  "model.decoder.layers.26.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
502
  "model.decoder.layers.26.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
503
  "model.decoder.layers.26.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
@@ -522,7 +502,6 @@
522
  "model.decoder.layers.27.fc2.weight": "model-00002-of-00002.safetensors",
523
  "model.decoder.layers.27.final_layer_norm.bias": "model-00002-of-00002.safetensors",
524
  "model.decoder.layers.27.final_layer_norm.weight": "model-00002-of-00002.safetensors",
525
- "model.decoder.layers.27.no_speech_bias": "model-00002-of-00002.safetensors",
526
  "model.decoder.layers.27.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
527
  "model.decoder.layers.27.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
528
  "model.decoder.layers.27.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
@@ -547,7 +526,6 @@
547
  "model.decoder.layers.28.fc2.weight": "model-00002-of-00002.safetensors",
548
  "model.decoder.layers.28.final_layer_norm.bias": "model-00002-of-00002.safetensors",
549
  "model.decoder.layers.28.final_layer_norm.weight": "model-00002-of-00002.safetensors",
550
- "model.decoder.layers.28.no_speech_bias": "model-00002-of-00002.safetensors",
551
  "model.decoder.layers.28.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
552
  "model.decoder.layers.28.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
553
  "model.decoder.layers.28.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
@@ -572,7 +550,6 @@
572
  "model.decoder.layers.29.fc2.weight": "model-00002-of-00002.safetensors",
573
  "model.decoder.layers.29.final_layer_norm.bias": "model-00002-of-00002.safetensors",
574
  "model.decoder.layers.29.final_layer_norm.weight": "model-00002-of-00002.safetensors",
575
- "model.decoder.layers.29.no_speech_bias": "model-00002-of-00002.safetensors",
576
  "model.decoder.layers.29.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
577
  "model.decoder.layers.29.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
578
  "model.decoder.layers.29.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
@@ -597,7 +574,6 @@
597
  "model.decoder.layers.3.fc2.weight": "model-00001-of-00002.safetensors",
598
  "model.decoder.layers.3.final_layer_norm.bias": "model-00001-of-00002.safetensors",
599
  "model.decoder.layers.3.final_layer_norm.weight": "model-00001-of-00002.safetensors",
600
- "model.decoder.layers.3.no_speech_bias": "model-00001-of-00002.safetensors",
601
  "model.decoder.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
602
  "model.decoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
603
  "model.decoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -622,7 +598,6 @@
622
  "model.decoder.layers.30.fc2.weight": "model-00002-of-00002.safetensors",
623
  "model.decoder.layers.30.final_layer_norm.bias": "model-00002-of-00002.safetensors",
624
  "model.decoder.layers.30.final_layer_norm.weight": "model-00002-of-00002.safetensors",
625
- "model.decoder.layers.30.no_speech_bias": "model-00002-of-00002.safetensors",
626
  "model.decoder.layers.30.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
627
  "model.decoder.layers.30.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
628
  "model.decoder.layers.30.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
@@ -647,7 +622,6 @@
647
  "model.decoder.layers.31.fc2.weight": "model-00002-of-00002.safetensors",
648
  "model.decoder.layers.31.final_layer_norm.bias": "model-00002-of-00002.safetensors",
649
  "model.decoder.layers.31.final_layer_norm.weight": "model-00002-of-00002.safetensors",
650
- "model.decoder.layers.31.no_speech_bias": "model-00002-of-00002.safetensors",
651
  "model.decoder.layers.31.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
652
  "model.decoder.layers.31.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
653
  "model.decoder.layers.31.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
@@ -672,7 +646,6 @@
672
  "model.decoder.layers.4.fc2.weight": "model-00001-of-00002.safetensors",
673
  "model.decoder.layers.4.final_layer_norm.bias": "model-00001-of-00002.safetensors",
674
  "model.decoder.layers.4.final_layer_norm.weight": "model-00001-of-00002.safetensors",
675
- "model.decoder.layers.4.no_speech_bias": "model-00001-of-00002.safetensors",
676
  "model.decoder.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
677
  "model.decoder.layers.4.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
678
  "model.decoder.layers.4.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -697,7 +670,6 @@
697
  "model.decoder.layers.5.fc2.weight": "model-00001-of-00002.safetensors",
698
  "model.decoder.layers.5.final_layer_norm.bias": "model-00001-of-00002.safetensors",
699
  "model.decoder.layers.5.final_layer_norm.weight": "model-00001-of-00002.safetensors",
700
- "model.decoder.layers.5.no_speech_bias": "model-00001-of-00002.safetensors",
701
  "model.decoder.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
702
  "model.decoder.layers.5.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
703
  "model.decoder.layers.5.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -722,7 +694,6 @@
722
  "model.decoder.layers.6.fc2.weight": "model-00001-of-00002.safetensors",
723
  "model.decoder.layers.6.final_layer_norm.bias": "model-00001-of-00002.safetensors",
724
  "model.decoder.layers.6.final_layer_norm.weight": "model-00001-of-00002.safetensors",
725
- "model.decoder.layers.6.no_speech_bias": "model-00001-of-00002.safetensors",
726
  "model.decoder.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
727
  "model.decoder.layers.6.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
728
  "model.decoder.layers.6.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -747,7 +718,6 @@
747
  "model.decoder.layers.7.fc2.weight": "model-00001-of-00002.safetensors",
748
  "model.decoder.layers.7.final_layer_norm.bias": "model-00001-of-00002.safetensors",
749
  "model.decoder.layers.7.final_layer_norm.weight": "model-00001-of-00002.safetensors",
750
- "model.decoder.layers.7.no_speech_bias": "model-00001-of-00002.safetensors",
751
  "model.decoder.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
752
  "model.decoder.layers.7.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
753
  "model.decoder.layers.7.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -772,7 +742,6 @@
772
  "model.decoder.layers.8.fc2.weight": "model-00001-of-00002.safetensors",
773
  "model.decoder.layers.8.final_layer_norm.bias": "model-00001-of-00002.safetensors",
774
  "model.decoder.layers.8.final_layer_norm.weight": "model-00001-of-00002.safetensors",
775
- "model.decoder.layers.8.no_speech_bias": "model-00001-of-00002.safetensors",
776
  "model.decoder.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
777
  "model.decoder.layers.8.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
778
  "model.decoder.layers.8.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
@@ -797,7 +766,6 @@
797
  "model.decoder.layers.9.fc2.weight": "model-00001-of-00002.safetensors",
798
  "model.decoder.layers.9.final_layer_norm.bias": "model-00001-of-00002.safetensors",
799
  "model.decoder.layers.9.final_layer_norm.weight": "model-00001-of-00002.safetensors",
800
- "model.decoder.layers.9.no_speech_bias": "model-00001-of-00002.safetensors",
801
  "model.decoder.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
802
  "model.decoder.layers.9.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
803
  "model.decoder.layers.9.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 6173219840
4
  },
5
  "weight_map": {
6
  "model.decoder.embed_positions.weight": "model-00001-of-00002.safetensors",
7
  "model.decoder.embed_tokens.weight": "model-00001-of-00002.safetensors",
8
+ "model.decoder.layer_norm.bias": "model-00002-of-00002.safetensors",
9
+ "model.decoder.layer_norm.weight": "model-00002-of-00002.safetensors",
10
  "model.decoder.layers.0.encoder_attn.k_proj.weight": "model-00001-of-00002.safetensors",
11
  "model.decoder.layers.0.encoder_attn.out_proj.bias": "model-00001-of-00002.safetensors",
12
  "model.decoder.layers.0.encoder_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
22
  "model.decoder.layers.0.fc2.weight": "model-00001-of-00002.safetensors",
23
  "model.decoder.layers.0.final_layer_norm.bias": "model-00001-of-00002.safetensors",
24
  "model.decoder.layers.0.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
25
  "model.decoder.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
26
  "model.decoder.layers.0.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
27
  "model.decoder.layers.0.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
46
  "model.decoder.layers.1.fc2.weight": "model-00001-of-00002.safetensors",
47
  "model.decoder.layers.1.final_layer_norm.bias": "model-00001-of-00002.safetensors",
48
  "model.decoder.layers.1.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
49
  "model.decoder.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
50
  "model.decoder.layers.1.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
51
  "model.decoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
70
  "model.decoder.layers.10.fc2.weight": "model-00001-of-00002.safetensors",
71
  "model.decoder.layers.10.final_layer_norm.bias": "model-00001-of-00002.safetensors",
72
  "model.decoder.layers.10.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
73
  "model.decoder.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
74
  "model.decoder.layers.10.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
75
  "model.decoder.layers.10.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
94
  "model.decoder.layers.11.fc2.weight": "model-00001-of-00002.safetensors",
95
  "model.decoder.layers.11.final_layer_norm.bias": "model-00001-of-00002.safetensors",
96
  "model.decoder.layers.11.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
97
  "model.decoder.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
98
  "model.decoder.layers.11.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
99
  "model.decoder.layers.11.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
118
  "model.decoder.layers.12.fc2.weight": "model-00001-of-00002.safetensors",
119
  "model.decoder.layers.12.final_layer_norm.bias": "model-00001-of-00002.safetensors",
120
  "model.decoder.layers.12.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
121
  "model.decoder.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
122
  "model.decoder.layers.12.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
123
  "model.decoder.layers.12.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
142
  "model.decoder.layers.13.fc2.weight": "model-00001-of-00002.safetensors",
143
  "model.decoder.layers.13.final_layer_norm.bias": "model-00001-of-00002.safetensors",
144
  "model.decoder.layers.13.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
145
  "model.decoder.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
146
  "model.decoder.layers.13.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
147
  "model.decoder.layers.13.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
166
  "model.decoder.layers.14.fc2.weight": "model-00001-of-00002.safetensors",
167
  "model.decoder.layers.14.final_layer_norm.bias": "model-00001-of-00002.safetensors",
168
  "model.decoder.layers.14.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
169
  "model.decoder.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
170
  "model.decoder.layers.14.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
171
  "model.decoder.layers.14.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
190
  "model.decoder.layers.15.fc2.weight": "model-00001-of-00002.safetensors",
191
  "model.decoder.layers.15.final_layer_norm.bias": "model-00001-of-00002.safetensors",
192
  "model.decoder.layers.15.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
193
  "model.decoder.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
194
  "model.decoder.layers.15.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
195
  "model.decoder.layers.15.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
214
  "model.decoder.layers.16.fc2.weight": "model-00001-of-00002.safetensors",
215
  "model.decoder.layers.16.final_layer_norm.bias": "model-00001-of-00002.safetensors",
216
  "model.decoder.layers.16.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
217
  "model.decoder.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
218
  "model.decoder.layers.16.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
219
  "model.decoder.layers.16.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
238
  "model.decoder.layers.17.fc2.weight": "model-00001-of-00002.safetensors",
239
  "model.decoder.layers.17.final_layer_norm.bias": "model-00001-of-00002.safetensors",
240
  "model.decoder.layers.17.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
241
  "model.decoder.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
242
  "model.decoder.layers.17.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
243
  "model.decoder.layers.17.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
262
  "model.decoder.layers.18.fc2.weight": "model-00001-of-00002.safetensors",
263
  "model.decoder.layers.18.final_layer_norm.bias": "model-00001-of-00002.safetensors",
264
  "model.decoder.layers.18.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
265
  "model.decoder.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
266
  "model.decoder.layers.18.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
267
  "model.decoder.layers.18.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
286
  "model.decoder.layers.19.fc2.weight": "model-00001-of-00002.safetensors",
287
  "model.decoder.layers.19.final_layer_norm.bias": "model-00001-of-00002.safetensors",
288
  "model.decoder.layers.19.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
289
  "model.decoder.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
290
  "model.decoder.layers.19.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
291
  "model.decoder.layers.19.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
310
  "model.decoder.layers.2.fc2.weight": "model-00001-of-00002.safetensors",
311
  "model.decoder.layers.2.final_layer_norm.bias": "model-00001-of-00002.safetensors",
312
  "model.decoder.layers.2.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
313
  "model.decoder.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
314
  "model.decoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
315
  "model.decoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
334
  "model.decoder.layers.20.fc2.weight": "model-00002-of-00002.safetensors",
335
  "model.decoder.layers.20.final_layer_norm.bias": "model-00002-of-00002.safetensors",
336
  "model.decoder.layers.20.final_layer_norm.weight": "model-00002-of-00002.safetensors",
 
337
  "model.decoder.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
338
  "model.decoder.layers.20.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
339
  "model.decoder.layers.20.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
358
  "model.decoder.layers.21.fc2.weight": "model-00002-of-00002.safetensors",
359
  "model.decoder.layers.21.final_layer_norm.bias": "model-00002-of-00002.safetensors",
360
  "model.decoder.layers.21.final_layer_norm.weight": "model-00002-of-00002.safetensors",
 
361
  "model.decoder.layers.21.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
362
  "model.decoder.layers.21.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
363
  "model.decoder.layers.21.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
 
382
  "model.decoder.layers.22.fc2.weight": "model-00002-of-00002.safetensors",
383
  "model.decoder.layers.22.final_layer_norm.bias": "model-00002-of-00002.safetensors",
384
  "model.decoder.layers.22.final_layer_norm.weight": "model-00002-of-00002.safetensors",
 
385
  "model.decoder.layers.22.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
386
  "model.decoder.layers.22.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
387
  "model.decoder.layers.22.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
 
406
  "model.decoder.layers.23.fc2.weight": "model-00002-of-00002.safetensors",
407
  "model.decoder.layers.23.final_layer_norm.bias": "model-00002-of-00002.safetensors",
408
  "model.decoder.layers.23.final_layer_norm.weight": "model-00002-of-00002.safetensors",
 
409
  "model.decoder.layers.23.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
410
  "model.decoder.layers.23.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
411
  "model.decoder.layers.23.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
 
430
  "model.decoder.layers.24.fc2.weight": "model-00002-of-00002.safetensors",
431
  "model.decoder.layers.24.final_layer_norm.bias": "model-00002-of-00002.safetensors",
432
  "model.decoder.layers.24.final_layer_norm.weight": "model-00002-of-00002.safetensors",
 
433
  "model.decoder.layers.24.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
434
  "model.decoder.layers.24.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
435
  "model.decoder.layers.24.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
 
454
  "model.decoder.layers.25.fc2.weight": "model-00002-of-00002.safetensors",
455
  "model.decoder.layers.25.final_layer_norm.bias": "model-00002-of-00002.safetensors",
456
  "model.decoder.layers.25.final_layer_norm.weight": "model-00002-of-00002.safetensors",
 
457
  "model.decoder.layers.25.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
458
  "model.decoder.layers.25.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
459
  "model.decoder.layers.25.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
 
478
  "model.decoder.layers.26.fc2.weight": "model-00002-of-00002.safetensors",
479
  "model.decoder.layers.26.final_layer_norm.bias": "model-00002-of-00002.safetensors",
480
  "model.decoder.layers.26.final_layer_norm.weight": "model-00002-of-00002.safetensors",
 
481
  "model.decoder.layers.26.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
482
  "model.decoder.layers.26.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
483
  "model.decoder.layers.26.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
 
502
  "model.decoder.layers.27.fc2.weight": "model-00002-of-00002.safetensors",
503
  "model.decoder.layers.27.final_layer_norm.bias": "model-00002-of-00002.safetensors",
504
  "model.decoder.layers.27.final_layer_norm.weight": "model-00002-of-00002.safetensors",
 
505
  "model.decoder.layers.27.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
506
  "model.decoder.layers.27.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
507
  "model.decoder.layers.27.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
 
526
  "model.decoder.layers.28.fc2.weight": "model-00002-of-00002.safetensors",
527
  "model.decoder.layers.28.final_layer_norm.bias": "model-00002-of-00002.safetensors",
528
  "model.decoder.layers.28.final_layer_norm.weight": "model-00002-of-00002.safetensors",
 
529
  "model.decoder.layers.28.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
530
  "model.decoder.layers.28.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
531
  "model.decoder.layers.28.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
 
550
  "model.decoder.layers.29.fc2.weight": "model-00002-of-00002.safetensors",
551
  "model.decoder.layers.29.final_layer_norm.bias": "model-00002-of-00002.safetensors",
552
  "model.decoder.layers.29.final_layer_norm.weight": "model-00002-of-00002.safetensors",
 
553
  "model.decoder.layers.29.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
554
  "model.decoder.layers.29.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
555
  "model.decoder.layers.29.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
 
574
  "model.decoder.layers.3.fc2.weight": "model-00001-of-00002.safetensors",
575
  "model.decoder.layers.3.final_layer_norm.bias": "model-00001-of-00002.safetensors",
576
  "model.decoder.layers.3.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
577
  "model.decoder.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
578
  "model.decoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
579
  "model.decoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
598
  "model.decoder.layers.30.fc2.weight": "model-00002-of-00002.safetensors",
599
  "model.decoder.layers.30.final_layer_norm.bias": "model-00002-of-00002.safetensors",
600
  "model.decoder.layers.30.final_layer_norm.weight": "model-00002-of-00002.safetensors",
 
601
  "model.decoder.layers.30.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
602
  "model.decoder.layers.30.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
603
  "model.decoder.layers.30.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
 
622
  "model.decoder.layers.31.fc2.weight": "model-00002-of-00002.safetensors",
623
  "model.decoder.layers.31.final_layer_norm.bias": "model-00002-of-00002.safetensors",
624
  "model.decoder.layers.31.final_layer_norm.weight": "model-00002-of-00002.safetensors",
 
625
  "model.decoder.layers.31.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
626
  "model.decoder.layers.31.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
627
  "model.decoder.layers.31.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
 
646
  "model.decoder.layers.4.fc2.weight": "model-00001-of-00002.safetensors",
647
  "model.decoder.layers.4.final_layer_norm.bias": "model-00001-of-00002.safetensors",
648
  "model.decoder.layers.4.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
649
  "model.decoder.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
650
  "model.decoder.layers.4.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
651
  "model.decoder.layers.4.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
670
  "model.decoder.layers.5.fc2.weight": "model-00001-of-00002.safetensors",
671
  "model.decoder.layers.5.final_layer_norm.bias": "model-00001-of-00002.safetensors",
672
  "model.decoder.layers.5.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
673
  "model.decoder.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
674
  "model.decoder.layers.5.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
675
  "model.decoder.layers.5.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
694
  "model.decoder.layers.6.fc2.weight": "model-00001-of-00002.safetensors",
695
  "model.decoder.layers.6.final_layer_norm.bias": "model-00001-of-00002.safetensors",
696
  "model.decoder.layers.6.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
697
  "model.decoder.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
698
  "model.decoder.layers.6.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
699
  "model.decoder.layers.6.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
718
  "model.decoder.layers.7.fc2.weight": "model-00001-of-00002.safetensors",
719
  "model.decoder.layers.7.final_layer_norm.bias": "model-00001-of-00002.safetensors",
720
  "model.decoder.layers.7.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
721
  "model.decoder.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
722
  "model.decoder.layers.7.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
723
  "model.decoder.layers.7.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
742
  "model.decoder.layers.8.fc2.weight": "model-00001-of-00002.safetensors",
743
  "model.decoder.layers.8.final_layer_norm.bias": "model-00001-of-00002.safetensors",
744
  "model.decoder.layers.8.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
745
  "model.decoder.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
746
  "model.decoder.layers.8.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
747
  "model.decoder.layers.8.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
 
766
  "model.decoder.layers.9.fc2.weight": "model-00001-of-00002.safetensors",
767
  "model.decoder.layers.9.final_layer_norm.bias": "model-00001-of-00002.safetensors",
768
  "model.decoder.layers.9.final_layer_norm.weight": "model-00001-of-00002.safetensors",
 
769
  "model.decoder.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
770
  "model.decoder.layers.9.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
771
  "model.decoder.layers.9.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",