NeMo
nvidia
jiaqiz commited on
Commit
80e399c
1 Parent(s): c164158

Add files using large-upload tool

Browse files
Files changed (26) hide show
  1. .gitattributes +25 -0
  2. model_weights/model.decoder.layers.self_attention.linear_proj.weight/12.0.3 +3 -0
  3. model_weights/model.decoder.layers.self_attention.linear_proj.weight/14.0.3 +3 -0
  4. model_weights/model.decoder.layers.self_attention.linear_proj.weight/16.0.7 +3 -0
  5. model_weights/model.decoder.layers.self_attention.linear_proj.weight/17.0.4 +3 -0
  6. model_weights/model.decoder.layers.self_attention.linear_proj.weight/2.0.5 +3 -0
  7. model_weights/model.decoder.layers.self_attention.linear_proj.weight/24.0.1 +3 -0
  8. model_weights/model.decoder.layers.self_attention.linear_proj.weight/25.0.1 +3 -0
  9. model_weights/model.decoder.layers.self_attention.linear_proj.weight/29.0.0 +3 -0
  10. model_weights/model.decoder.layers.self_attention.linear_proj.weight/32.0.6 +3 -0
  11. model_weights/model.decoder.layers.self_attention.linear_proj.weight/36.0.5 +3 -0
  12. model_weights/model.decoder.layers.self_attention.linear_proj.weight/4.0.7 +3 -0
  13. model_weights/model.decoder.layers.self_attention.linear_proj.weight/40.0.2 +3 -0
  14. model_weights/model.decoder.layers.self_attention.linear_proj.weight/49.0.2 +3 -0
  15. model_weights/model.decoder.layers.self_attention.linear_proj.weight/49.0.7 +3 -0
  16. model_weights/model.decoder.layers.self_attention.linear_proj.weight/52.0.7 +3 -0
  17. model_weights/model.decoder.layers.self_attention.linear_proj.weight/53.0.5 +3 -0
  18. model_weights/model.decoder.layers.self_attention.linear_proj.weight/54.0.4 +3 -0
  19. model_weights/model.decoder.layers.self_attention.linear_proj.weight/65.0.5 +3 -0
  20. model_weights/model.decoder.layers.self_attention.linear_proj.weight/66.0.5 +3 -0
  21. model_weights/model.decoder.layers.self_attention.linear_proj.weight/7.0.3 +3 -0
  22. model_weights/model.decoder.layers.self_attention.linear_proj.weight/80.0.2 +3 -0
  23. model_weights/model.decoder.layers.self_attention.linear_proj.weight/80.0.5 +3 -0
  24. model_weights/model.decoder.layers.self_attention.linear_proj.weight/84.0.5 +3 -0
  25. model_weights/model.decoder.layers.self_attention.linear_proj.weight/85.0.0 +3 -0
  26. model_weights/model.decoder.layers.self_attention.linear_proj.weight/89.0.2 +3 -0
.gitattributes CHANGED
@@ -2545,3 +2545,28 @@ model_weights/model.decoder.layers.self_attention.linear_proj.weight/47.0.3 filt
2545
  model_weights/model.decoder.layers.self_attention.linear_proj.weight/39.0.4 filter=lfs diff=lfs merge=lfs -text
2546
  model_weights/model.decoder.layers.self_attention.linear_proj.weight/76.0.3 filter=lfs diff=lfs merge=lfs -text
2547
  model_weights/model.decoder.layers.self_attention.linear_proj.weight/64.0.0 filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2545
  model_weights/model.decoder.layers.self_attention.linear_proj.weight/39.0.4 filter=lfs diff=lfs merge=lfs -text
2546
  model_weights/model.decoder.layers.self_attention.linear_proj.weight/76.0.3 filter=lfs diff=lfs merge=lfs -text
2547
  model_weights/model.decoder.layers.self_attention.linear_proj.weight/64.0.0 filter=lfs diff=lfs merge=lfs -text
2548
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/40.0.2 filter=lfs diff=lfs merge=lfs -text
2549
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/80.0.5 filter=lfs diff=lfs merge=lfs -text
2550
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/29.0.0 filter=lfs diff=lfs merge=lfs -text
2551
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/24.0.1 filter=lfs diff=lfs merge=lfs -text
2552
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/49.0.7 filter=lfs diff=lfs merge=lfs -text
2553
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/65.0.5 filter=lfs diff=lfs merge=lfs -text
2554
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/52.0.7 filter=lfs diff=lfs merge=lfs -text
2555
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/80.0.2 filter=lfs diff=lfs merge=lfs -text
2556
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/32.0.6 filter=lfs diff=lfs merge=lfs -text
2557
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/54.0.4 filter=lfs diff=lfs merge=lfs -text
2558
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/2.0.5 filter=lfs diff=lfs merge=lfs -text
2559
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/12.0.3 filter=lfs diff=lfs merge=lfs -text
2560
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/49.0.2 filter=lfs diff=lfs merge=lfs -text
2561
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/84.0.5 filter=lfs diff=lfs merge=lfs -text
2562
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/85.0.0 filter=lfs diff=lfs merge=lfs -text
2563
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/14.0.3 filter=lfs diff=lfs merge=lfs -text
2564
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/89.0.2 filter=lfs diff=lfs merge=lfs -text
2565
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/66.0.5 filter=lfs diff=lfs merge=lfs -text
2566
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/25.0.1 filter=lfs diff=lfs merge=lfs -text
2567
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/16.0.7 filter=lfs diff=lfs merge=lfs -text
2568
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/17.0.4 filter=lfs diff=lfs merge=lfs -text
2569
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/7.0.3 filter=lfs diff=lfs merge=lfs -text
2570
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/36.0.5 filter=lfs diff=lfs merge=lfs -text
2571
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/53.0.5 filter=lfs diff=lfs merge=lfs -text
2572
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/4.0.7 filter=lfs diff=lfs merge=lfs -text
model_weights/model.decoder.layers.self_attention.linear_proj.weight/12.0.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15a56fc51a11c38d85af2d9452249ba32379f102e547dacf9fe96a3268986dfe
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/14.0.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dffb0b4e00b8c9e1e730ae692707787bc026b18d1bfe64fd2db96a809da15b69
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/16.0.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d460ff00c8756cdb7e1e2ded9b0c35e5b4c9ea8c66bd8d68e42cf2e5906435b4
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/17.0.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8093d6263fd627e547fca9caeeae20be782a5d0106312eead112a3f51158c02
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/2.0.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:834f21a2c0fd6c7fc3eb34cc978d926d30dfed842824222e9de616a40043aa4c
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/24.0.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0c31202b26d567115905dd9b6ee88ae26baab194581ee86a8b56028812668b5
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/25.0.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de7ee44931d761fd850d1b93320019d150ee4ea4754342ebee2e443a3912aef2
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/29.0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4e0b4023d6d8a67a594adfdadf40704fae47c15043eb8d3e591d6ca68b5c001
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/32.0.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51ae93b3a9967cf229fec8eca2a982d4f22d7db5599e3a54ef19bb9b8fd75555
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/36.0.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f206ce50c1a58065df6d976305a692fbd14fdfb35d9460279e368d7095346a34
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/4.0.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d77c22063fc72352fd72c549dd44ae7fa85f6efcf8cdc13e80803579ea0817d
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/40.0.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:492d0c5fd0a2077ec3f06c496321a629467671c8ec3e9db5fd702b73c054bfdc
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/49.0.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6160cd2559a4b4092937e73fceb6ceb46e475530f9f6ade22b9292d960b2be0
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/49.0.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c186c7d0cc6cbbbaeae140c7134fc27a7fae56602161ca587a7e5c796499a5f
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/52.0.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce3c0921cdcd84c20f5f95583bf15918a80d8dc6f3b869e2255b73cc4380172a
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/53.0.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db9b8fc3dd1b23b25d2c6f6201dc029fc99622223ff3ad52b57d1f37556bf8e3
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/54.0.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f32b449e94491b50562cdcde417f455aa68592d1abedb6f163efb2c31a30dfa7
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/65.0.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:599b18dfa69457bb133541862cf3aa996ec0d28958206b3135562e9326c5dd82
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/66.0.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cced0d83d6fd9c11e5d52efd79f6e082359d5fc640a841cdbb0451a2ffc4287c
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/7.0.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80bc1bede7d065e04f4fba2bcfbbea46bebfd560918c23440f754747436cea36
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/80.0.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64d66a44be8d55da4543c82591cdd06fcf95178b4a5b407b5ad4d0114c3b936e
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/80.0.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:657666afcc7513a7dc284f2f372d2712acd7a78d40d881dda78fedb55756f26b
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/84.0.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f47409fa2b99ceccc8e52dd98cdeb9fe36d51b3c2f4313f7c62f10482f4131db
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/85.0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b5ca3ccbec971cb2665174dd14679e51ce576c971fe5fc0350f88acb4bd3ab2
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/89.0.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43008d6f5bee8693f927559864442336768c342a1b2e660178274dbe0a675f6b
3
+ size 84934656