lhallee commited on
Commit
447e27f
1 Parent(s): 3170072

Upload ProteinVec

Browse files
config.json CHANGED
@@ -105,7 +105,7 @@
105
  "transformers_version": "4.38.1",
106
  "use_cache": true,
107
  "vec_activation": "relu",
108
- "vec_d_model": 1024,
109
  "vec_dim_feedforward": 2048,
110
  "vec_dropout": 0.1,
111
  "vec_lr0": 0.0001,
 
105
  "transformers_version": "4.38.1",
106
  "use_cache": true,
107
  "vec_activation": "relu",
108
+ "vec_d_model": 512,
109
  "vec_dim_feedforward": 2048,
110
  "vec_dropout": 0.1,
111
  "vec_lr0": 0.0001,
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac4ed24c74dce0afa06faf9efdd2e32fa179002491ddc2e8eeac83cb9e93403f
3
- size 4995336712
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a2e2a081a78c4d5ce1d72a5e545e211aa6a83140cd6954b097ceff1a2e52d49
3
+ size 4996418440
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8753bd949ac67f613ff84915e80694d153a9dcae910c3b587522a20360baadc7
3
- size 667829856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:365bbe7727ae454be8bb8fdb173a49f0ca762a3afb3b4b6a7055a02f45eb92b3
3
+ size 623719592
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 5663102976
4
  },
5
  "weight_map": {
6
  "moe.encoder.layers.0.linear1.bias": "model-00001-of-00002.safetensors",
@@ -297,30 +297,30 @@
297
  "moe.model_aspect_tmvec.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
298
  "moe.model_aspect_tmvec.encoder.layers.2.linear1.bias": "model-00001-of-00002.safetensors",
299
  "moe.model_aspect_tmvec.encoder.layers.2.linear1.weight": "model-00001-of-00002.safetensors",
300
- "moe.model_aspect_tmvec.encoder.layers.2.linear2.bias": "model-00002-of-00002.safetensors",
301
- "moe.model_aspect_tmvec.encoder.layers.2.linear2.weight": "model-00002-of-00002.safetensors",
302
- "moe.model_aspect_tmvec.encoder.layers.2.norm1.bias": "model-00002-of-00002.safetensors",
303
- "moe.model_aspect_tmvec.encoder.layers.2.norm1.weight": "model-00002-of-00002.safetensors",
304
- "moe.model_aspect_tmvec.encoder.layers.2.norm2.bias": "model-00002-of-00002.safetensors",
305
- "moe.model_aspect_tmvec.encoder.layers.2.norm2.weight": "model-00002-of-00002.safetensors",
306
  "moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_bias": "model-00001-of-00002.safetensors",
307
  "moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_weight": "model-00001-of-00002.safetensors",
308
  "moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
309
  "moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
310
- "moe.model_aspect_tmvec.encoder.layers.3.linear1.bias": "model-00002-of-00002.safetensors",
311
- "moe.model_aspect_tmvec.encoder.layers.3.linear1.weight": "model-00002-of-00002.safetensors",
312
- "moe.model_aspect_tmvec.encoder.layers.3.linear2.bias": "model-00002-of-00002.safetensors",
313
- "moe.model_aspect_tmvec.encoder.layers.3.linear2.weight": "model-00002-of-00002.safetensors",
314
- "moe.model_aspect_tmvec.encoder.layers.3.norm1.bias": "model-00002-of-00002.safetensors",
315
- "moe.model_aspect_tmvec.encoder.layers.3.norm1.weight": "model-00002-of-00002.safetensors",
316
- "moe.model_aspect_tmvec.encoder.layers.3.norm2.bias": "model-00002-of-00002.safetensors",
317
- "moe.model_aspect_tmvec.encoder.layers.3.norm2.weight": "model-00002-of-00002.safetensors",
318
- "moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_bias": "model-00002-of-00002.safetensors",
319
- "moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_weight": "model-00002-of-00002.safetensors",
320
- "moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
321
- "moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
322
- "moe.model_aspect_tmvec.mlp.bias": "model-00002-of-00002.safetensors",
323
- "moe.model_aspect_tmvec.mlp.weight": "model-00002-of-00002.safetensors",
324
  "t5.encoder.block.0.layer.0.SelfAttention.k.weight": "model-00001-of-00002.safetensors",
325
  "t5.encoder.block.0.layer.0.SelfAttention.o.weight": "model-00001-of-00002.safetensors",
326
  "t5.encoder.block.0.layer.0.SelfAttention.q.weight": "model-00001-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 5620074496
4
  },
5
  "weight_map": {
6
  "moe.encoder.layers.0.linear1.bias": "model-00001-of-00002.safetensors",
 
297
  "moe.model_aspect_tmvec.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
298
  "moe.model_aspect_tmvec.encoder.layers.2.linear1.bias": "model-00001-of-00002.safetensors",
299
  "moe.model_aspect_tmvec.encoder.layers.2.linear1.weight": "model-00001-of-00002.safetensors",
300
+ "moe.model_aspect_tmvec.encoder.layers.2.linear2.bias": "model-00001-of-00002.safetensors",
301
+ "moe.model_aspect_tmvec.encoder.layers.2.linear2.weight": "model-00001-of-00002.safetensors",
302
+ "moe.model_aspect_tmvec.encoder.layers.2.norm1.bias": "model-00001-of-00002.safetensors",
303
+ "moe.model_aspect_tmvec.encoder.layers.2.norm1.weight": "model-00001-of-00002.safetensors",
304
+ "moe.model_aspect_tmvec.encoder.layers.2.norm2.bias": "model-00001-of-00002.safetensors",
305
+ "moe.model_aspect_tmvec.encoder.layers.2.norm2.weight": "model-00001-of-00002.safetensors",
306
  "moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_bias": "model-00001-of-00002.safetensors",
307
  "moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_weight": "model-00001-of-00002.safetensors",
308
  "moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
309
  "moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
310
+ "moe.model_aspect_tmvec.encoder.layers.3.linear1.bias": "model-00001-of-00002.safetensors",
311
+ "moe.model_aspect_tmvec.encoder.layers.3.linear1.weight": "model-00001-of-00002.safetensors",
312
+ "moe.model_aspect_tmvec.encoder.layers.3.linear2.bias": "model-00001-of-00002.safetensors",
313
+ "moe.model_aspect_tmvec.encoder.layers.3.linear2.weight": "model-00001-of-00002.safetensors",
314
+ "moe.model_aspect_tmvec.encoder.layers.3.norm1.bias": "model-00001-of-00002.safetensors",
315
+ "moe.model_aspect_tmvec.encoder.layers.3.norm1.weight": "model-00001-of-00002.safetensors",
316
+ "moe.model_aspect_tmvec.encoder.layers.3.norm2.bias": "model-00001-of-00002.safetensors",
317
+ "moe.model_aspect_tmvec.encoder.layers.3.norm2.weight": "model-00001-of-00002.safetensors",
318
+ "moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_bias": "model-00001-of-00002.safetensors",
319
+ "moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_weight": "model-00001-of-00002.safetensors",
320
+ "moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
321
+ "moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
322
+ "moe.model_aspect_tmvec.mlp.bias": "model-00001-of-00002.safetensors",
323
+ "moe.model_aspect_tmvec.mlp.weight": "model-00001-of-00002.safetensors",
324
  "t5.encoder.block.0.layer.0.SelfAttention.k.weight": "model-00001-of-00002.safetensors",
325
  "t5.encoder.block.0.layer.0.SelfAttention.o.weight": "model-00001-of-00002.safetensors",
326
  "t5.encoder.block.0.layer.0.SelfAttention.q.weight": "model-00001-of-00002.safetensors",