Upload ProteinVec
Browse files- config.json +1 -1
- model-00001-of-00002.safetensors +2 -2
- model-00002-of-00002.safetensors +2 -2
- model.safetensors.index.json +21 -21
config.json
CHANGED
@@ -105,7 +105,7 @@
|
|
105 |
"transformers_version": "4.38.1",
|
106 |
"use_cache": true,
|
107 |
"vec_activation": "relu",
|
108 |
-
"vec_d_model":
|
109 |
"vec_dim_feedforward": 2048,
|
110 |
"vec_dropout": 0.1,
|
111 |
"vec_lr0": 0.0001,
|
|
|
105 |
"transformers_version": "4.38.1",
|
106 |
"use_cache": true,
|
107 |
"vec_activation": "relu",
|
108 |
+
"vec_d_model": 512,
|
109 |
"vec_dim_feedforward": 2048,
|
110 |
"vec_dropout": 0.1,
|
111 |
"vec_lr0": 0.0001,
|
model-00001-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6a2e2a081a78c4d5ce1d72a5e545e211aa6a83140cd6954b097ceff1a2e52d49
|
3 |
+
size 4996418440
|
model-00002-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:365bbe7727ae454be8bb8fdb173a49f0ca762a3afb3b4b6a7055a02f45eb92b3
|
3 |
+
size 623719592
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"moe.encoder.layers.0.linear1.bias": "model-00001-of-00002.safetensors",
|
@@ -297,30 +297,30 @@
|
|
297 |
"moe.model_aspect_tmvec.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
298 |
"moe.model_aspect_tmvec.encoder.layers.2.linear1.bias": "model-00001-of-00002.safetensors",
|
299 |
"moe.model_aspect_tmvec.encoder.layers.2.linear1.weight": "model-00001-of-00002.safetensors",
|
300 |
-
"moe.model_aspect_tmvec.encoder.layers.2.linear2.bias": "model-
|
301 |
-
"moe.model_aspect_tmvec.encoder.layers.2.linear2.weight": "model-
|
302 |
-
"moe.model_aspect_tmvec.encoder.layers.2.norm1.bias": "model-
|
303 |
-
"moe.model_aspect_tmvec.encoder.layers.2.norm1.weight": "model-
|
304 |
-
"moe.model_aspect_tmvec.encoder.layers.2.norm2.bias": "model-
|
305 |
-
"moe.model_aspect_tmvec.encoder.layers.2.norm2.weight": "model-
|
306 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_bias": "model-00001-of-00002.safetensors",
|
307 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_weight": "model-00001-of-00002.safetensors",
|
308 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
309 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
310 |
-
"moe.model_aspect_tmvec.encoder.layers.3.linear1.bias": "model-
|
311 |
-
"moe.model_aspect_tmvec.encoder.layers.3.linear1.weight": "model-
|
312 |
-
"moe.model_aspect_tmvec.encoder.layers.3.linear2.bias": "model-
|
313 |
-
"moe.model_aspect_tmvec.encoder.layers.3.linear2.weight": "model-
|
314 |
-
"moe.model_aspect_tmvec.encoder.layers.3.norm1.bias": "model-
|
315 |
-
"moe.model_aspect_tmvec.encoder.layers.3.norm1.weight": "model-
|
316 |
-
"moe.model_aspect_tmvec.encoder.layers.3.norm2.bias": "model-
|
317 |
-
"moe.model_aspect_tmvec.encoder.layers.3.norm2.weight": "model-
|
318 |
-
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_bias": "model-
|
319 |
-
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_weight": "model-
|
320 |
-
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.bias": "model-
|
321 |
-
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.weight": "model-
|
322 |
-
"moe.model_aspect_tmvec.mlp.bias": "model-
|
323 |
-
"moe.model_aspect_tmvec.mlp.weight": "model-
|
324 |
"t5.encoder.block.0.layer.0.SelfAttention.k.weight": "model-00001-of-00002.safetensors",
|
325 |
"t5.encoder.block.0.layer.0.SelfAttention.o.weight": "model-00001-of-00002.safetensors",
|
326 |
"t5.encoder.block.0.layer.0.SelfAttention.q.weight": "model-00001-of-00002.safetensors",
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 5620074496
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"moe.encoder.layers.0.linear1.bias": "model-00001-of-00002.safetensors",
|
|
|
297 |
"moe.model_aspect_tmvec.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
298 |
"moe.model_aspect_tmvec.encoder.layers.2.linear1.bias": "model-00001-of-00002.safetensors",
|
299 |
"moe.model_aspect_tmvec.encoder.layers.2.linear1.weight": "model-00001-of-00002.safetensors",
|
300 |
+
"moe.model_aspect_tmvec.encoder.layers.2.linear2.bias": "model-00001-of-00002.safetensors",
|
301 |
+
"moe.model_aspect_tmvec.encoder.layers.2.linear2.weight": "model-00001-of-00002.safetensors",
|
302 |
+
"moe.model_aspect_tmvec.encoder.layers.2.norm1.bias": "model-00001-of-00002.safetensors",
|
303 |
+
"moe.model_aspect_tmvec.encoder.layers.2.norm1.weight": "model-00001-of-00002.safetensors",
|
304 |
+
"moe.model_aspect_tmvec.encoder.layers.2.norm2.bias": "model-00001-of-00002.safetensors",
|
305 |
+
"moe.model_aspect_tmvec.encoder.layers.2.norm2.weight": "model-00001-of-00002.safetensors",
|
306 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_bias": "model-00001-of-00002.safetensors",
|
307 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_weight": "model-00001-of-00002.safetensors",
|
308 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
309 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
310 |
+
"moe.model_aspect_tmvec.encoder.layers.3.linear1.bias": "model-00001-of-00002.safetensors",
|
311 |
+
"moe.model_aspect_tmvec.encoder.layers.3.linear1.weight": "model-00001-of-00002.safetensors",
|
312 |
+
"moe.model_aspect_tmvec.encoder.layers.3.linear2.bias": "model-00001-of-00002.safetensors",
|
313 |
+
"moe.model_aspect_tmvec.encoder.layers.3.linear2.weight": "model-00001-of-00002.safetensors",
|
314 |
+
"moe.model_aspect_tmvec.encoder.layers.3.norm1.bias": "model-00001-of-00002.safetensors",
|
315 |
+
"moe.model_aspect_tmvec.encoder.layers.3.norm1.weight": "model-00001-of-00002.safetensors",
|
316 |
+
"moe.model_aspect_tmvec.encoder.layers.3.norm2.bias": "model-00001-of-00002.safetensors",
|
317 |
+
"moe.model_aspect_tmvec.encoder.layers.3.norm2.weight": "model-00001-of-00002.safetensors",
|
318 |
+
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_bias": "model-00001-of-00002.safetensors",
|
319 |
+
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_weight": "model-00001-of-00002.safetensors",
|
320 |
+
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
321 |
+
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
322 |
+
"moe.model_aspect_tmvec.mlp.bias": "model-00001-of-00002.safetensors",
|
323 |
+
"moe.model_aspect_tmvec.mlp.weight": "model-00001-of-00002.safetensors",
|
324 |
"t5.encoder.block.0.layer.0.SelfAttention.k.weight": "model-00001-of-00002.safetensors",
|
325 |
"t5.encoder.block.0.layer.0.SelfAttention.o.weight": "model-00001-of-00002.safetensors",
|
326 |
"t5.encoder.block.0.layer.0.SelfAttention.q.weight": "model-00001-of-00002.safetensors",
|