Upload ProteinVec
Browse files- config.json +1 -1
- model-00001-of-00002.safetensors +2 -2
- model-00002-of-00002.safetensors +2 -2
- model.safetensors.index.json +21 -21
config.json
CHANGED
|
@@ -105,7 +105,7 @@
|
|
| 105 |
"transformers_version": "4.38.1",
|
| 106 |
"use_cache": true,
|
| 107 |
"vec_activation": "relu",
|
| 108 |
-
"vec_d_model":
|
| 109 |
"vec_dim_feedforward": 2048,
|
| 110 |
"vec_dropout": 0.1,
|
| 111 |
"vec_lr0": 0.0001,
|
|
|
|
| 105 |
"transformers_version": "4.38.1",
|
| 106 |
"use_cache": true,
|
| 107 |
"vec_activation": "relu",
|
| 108 |
+
"vec_d_model": 512,
|
| 109 |
"vec_dim_feedforward": 2048,
|
| 110 |
"vec_dropout": 0.1,
|
| 111 |
"vec_lr0": 0.0001,
|
model-00001-of-00002.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6a2e2a081a78c4d5ce1d72a5e545e211aa6a83140cd6954b097ceff1a2e52d49
|
| 3 |
+
size 4996418440
|
model-00002-of-00002.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:365bbe7727ae454be8bb8fdb173a49f0ca762a3afb3b4b6a7055a02f45eb92b3
|
| 3 |
+
size 623719592
|
model.safetensors.index.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
-
"total_size":
|
| 4 |
},
|
| 5 |
"weight_map": {
|
| 6 |
"moe.encoder.layers.0.linear1.bias": "model-00001-of-00002.safetensors",
|
|
@@ -297,30 +297,30 @@
|
|
| 297 |
"moe.model_aspect_tmvec.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
| 298 |
"moe.model_aspect_tmvec.encoder.layers.2.linear1.bias": "model-00001-of-00002.safetensors",
|
| 299 |
"moe.model_aspect_tmvec.encoder.layers.2.linear1.weight": "model-00001-of-00002.safetensors",
|
| 300 |
-
"moe.model_aspect_tmvec.encoder.layers.2.linear2.bias": "model-
|
| 301 |
-
"moe.model_aspect_tmvec.encoder.layers.2.linear2.weight": "model-
|
| 302 |
-
"moe.model_aspect_tmvec.encoder.layers.2.norm1.bias": "model-
|
| 303 |
-
"moe.model_aspect_tmvec.encoder.layers.2.norm1.weight": "model-
|
| 304 |
-
"moe.model_aspect_tmvec.encoder.layers.2.norm2.bias": "model-
|
| 305 |
-
"moe.model_aspect_tmvec.encoder.layers.2.norm2.weight": "model-
|
| 306 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_bias": "model-00001-of-00002.safetensors",
|
| 307 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_weight": "model-00001-of-00002.safetensors",
|
| 308 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
| 309 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
| 310 |
-
"moe.model_aspect_tmvec.encoder.layers.3.linear1.bias": "model-
|
| 311 |
-
"moe.model_aspect_tmvec.encoder.layers.3.linear1.weight": "model-
|
| 312 |
-
"moe.model_aspect_tmvec.encoder.layers.3.linear2.bias": "model-
|
| 313 |
-
"moe.model_aspect_tmvec.encoder.layers.3.linear2.weight": "model-
|
| 314 |
-
"moe.model_aspect_tmvec.encoder.layers.3.norm1.bias": "model-
|
| 315 |
-
"moe.model_aspect_tmvec.encoder.layers.3.norm1.weight": "model-
|
| 316 |
-
"moe.model_aspect_tmvec.encoder.layers.3.norm2.bias": "model-
|
| 317 |
-
"moe.model_aspect_tmvec.encoder.layers.3.norm2.weight": "model-
|
| 318 |
-
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_bias": "model-
|
| 319 |
-
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_weight": "model-
|
| 320 |
-
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.bias": "model-
|
| 321 |
-
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.weight": "model-
|
| 322 |
-
"moe.model_aspect_tmvec.mlp.bias": "model-
|
| 323 |
-
"moe.model_aspect_tmvec.mlp.weight": "model-
|
| 324 |
"t5.encoder.block.0.layer.0.SelfAttention.k.weight": "model-00001-of-00002.safetensors",
|
| 325 |
"t5.encoder.block.0.layer.0.SelfAttention.o.weight": "model-00001-of-00002.safetensors",
|
| 326 |
"t5.encoder.block.0.layer.0.SelfAttention.q.weight": "model-00001-of-00002.safetensors",
|
|
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
+
"total_size": 5620074496
|
| 4 |
},
|
| 5 |
"weight_map": {
|
| 6 |
"moe.encoder.layers.0.linear1.bias": "model-00001-of-00002.safetensors",
|
|
|
|
| 297 |
"moe.model_aspect_tmvec.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
| 298 |
"moe.model_aspect_tmvec.encoder.layers.2.linear1.bias": "model-00001-of-00002.safetensors",
|
| 299 |
"moe.model_aspect_tmvec.encoder.layers.2.linear1.weight": "model-00001-of-00002.safetensors",
|
| 300 |
+
"moe.model_aspect_tmvec.encoder.layers.2.linear2.bias": "model-00001-of-00002.safetensors",
|
| 301 |
+
"moe.model_aspect_tmvec.encoder.layers.2.linear2.weight": "model-00001-of-00002.safetensors",
|
| 302 |
+
"moe.model_aspect_tmvec.encoder.layers.2.norm1.bias": "model-00001-of-00002.safetensors",
|
| 303 |
+
"moe.model_aspect_tmvec.encoder.layers.2.norm1.weight": "model-00001-of-00002.safetensors",
|
| 304 |
+
"moe.model_aspect_tmvec.encoder.layers.2.norm2.bias": "model-00001-of-00002.safetensors",
|
| 305 |
+
"moe.model_aspect_tmvec.encoder.layers.2.norm2.weight": "model-00001-of-00002.safetensors",
|
| 306 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_bias": "model-00001-of-00002.safetensors",
|
| 307 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_weight": "model-00001-of-00002.safetensors",
|
| 308 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
| 309 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
| 310 |
+
"moe.model_aspect_tmvec.encoder.layers.3.linear1.bias": "model-00001-of-00002.safetensors",
|
| 311 |
+
"moe.model_aspect_tmvec.encoder.layers.3.linear1.weight": "model-00001-of-00002.safetensors",
|
| 312 |
+
"moe.model_aspect_tmvec.encoder.layers.3.linear2.bias": "model-00001-of-00002.safetensors",
|
| 313 |
+
"moe.model_aspect_tmvec.encoder.layers.3.linear2.weight": "model-00001-of-00002.safetensors",
|
| 314 |
+
"moe.model_aspect_tmvec.encoder.layers.3.norm1.bias": "model-00001-of-00002.safetensors",
|
| 315 |
+
"moe.model_aspect_tmvec.encoder.layers.3.norm1.weight": "model-00001-of-00002.safetensors",
|
| 316 |
+
"moe.model_aspect_tmvec.encoder.layers.3.norm2.bias": "model-00001-of-00002.safetensors",
|
| 317 |
+
"moe.model_aspect_tmvec.encoder.layers.3.norm2.weight": "model-00001-of-00002.safetensors",
|
| 318 |
+
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_bias": "model-00001-of-00002.safetensors",
|
| 319 |
+
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_weight": "model-00001-of-00002.safetensors",
|
| 320 |
+
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
| 321 |
+
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
| 322 |
+
"moe.model_aspect_tmvec.mlp.bias": "model-00001-of-00002.safetensors",
|
| 323 |
+
"moe.model_aspect_tmvec.mlp.weight": "model-00001-of-00002.safetensors",
|
| 324 |
"t5.encoder.block.0.layer.0.SelfAttention.k.weight": "model-00001-of-00002.safetensors",
|
| 325 |
"t5.encoder.block.0.layer.0.SelfAttention.o.weight": "model-00001-of-00002.safetensors",
|
| 326 |
"t5.encoder.block.0.layer.0.SelfAttention.q.weight": "model-00001-of-00002.safetensors",
|