patrickvonplaten commited on
Commit
ab890c9
·
1 Parent(s): b4c7653
Files changed (50) hide show
  1. text_encoder/config.json +19 -6
  2. text_encoder/model-00001-of-00023.safetensors +0 -0
  3. text_encoder/model-00002-of-00023.safetensors +0 -0
  4. text_encoder/model-00003-of-00023.safetensors +0 -0
  5. text_encoder/model-00004-of-00023.safetensors +0 -0
  6. text_encoder/model-00005-of-00023.safetensors +0 -0
  7. text_encoder/model-00006-of-00023.safetensors +0 -0
  8. text_encoder/model-00007-of-00023.safetensors +0 -0
  9. text_encoder/model-00008-of-00023.safetensors +0 -0
  10. text_encoder/model-00009-of-00023.safetensors +0 -0
  11. text_encoder/model-00010-of-00023.safetensors +0 -0
  12. text_encoder/model-00011-of-00023.safetensors +0 -0
  13. text_encoder/model-00012-of-00023.safetensors +0 -0
  14. text_encoder/model-00013-of-00023.safetensors +0 -0
  15. text_encoder/model-00014-of-00023.safetensors +0 -0
  16. text_encoder/model-00015-of-00023.safetensors +0 -0
  17. text_encoder/model-00016-of-00023.safetensors +0 -0
  18. text_encoder/model-00017-of-00023.safetensors +0 -0
  19. text_encoder/model-00018-of-00023.safetensors +0 -0
  20. text_encoder/model-00019-of-00023.safetensors +0 -0
  21. text_encoder/model-00020-of-00023.safetensors +0 -0
  22. text_encoder/model-00021-of-00023.safetensors +0 -0
  23. text_encoder/model-00022-of-00023.safetensors +0 -0
  24. text_encoder/model-00023-of-00023.safetensors +0 -0
  25. text_encoder/model.safetensors.index.json +50 -0
  26. text_encoder/pytorch_model-00001-of-00024.bin +3 -0
  27. text_encoder/pytorch_model-00002-of-00024.bin +3 -0
  28. text_encoder/pytorch_model-00003-of-00024.bin +3 -0
  29. text_encoder/pytorch_model-00004-of-00024.bin +3 -0
  30. text_encoder/pytorch_model-00005-of-00024.bin +3 -0
  31. text_encoder/pytorch_model-00006-of-00024.bin +3 -0
  32. text_encoder/pytorch_model-00007-of-00024.bin +3 -0
  33. text_encoder/pytorch_model-00008-of-00024.bin +3 -0
  34. text_encoder/pytorch_model-00009-of-00024.bin +3 -0
  35. text_encoder/pytorch_model-00010-of-00024.bin +3 -0
  36. text_encoder/pytorch_model-00011-of-00024.bin +3 -0
  37. text_encoder/pytorch_model-00012-of-00024.bin +3 -0
  38. text_encoder/pytorch_model-00013-of-00024.bin +3 -0
  39. text_encoder/pytorch_model-00014-of-00024.bin +3 -0
  40. text_encoder/pytorch_model-00015-of-00024.bin +3 -0
  41. text_encoder/pytorch_model-00016-of-00024.bin +3 -0
  42. text_encoder/pytorch_model-00017-of-00024.bin +3 -0
  43. text_encoder/pytorch_model-00018-of-00024.bin +3 -0
  44. text_encoder/pytorch_model-00019-of-00024.bin +3 -0
  45. text_encoder/pytorch_model-00020-of-00024.bin +3 -0
  46. text_encoder/pytorch_model-00021-of-00024.bin +3 -0
  47. text_encoder/pytorch_model-00022-of-00024.bin +3 -0
  48. text_encoder/pytorch_model-00023-of-00024.bin +3 -0
  49. text_encoder/pytorch_model-00024-of-00024.bin +3 -0
  50. text_encoder/pytorch_model.bin.index.json +51 -0
text_encoder/config.json CHANGED
@@ -1,23 +1,36 @@
1
  {
 
2
  "architectures": [
3
- "CLIPTextModel"
4
  ],
5
  "attention_dropout": 0.0,
6
  "bos_token_id": 0,
 
 
 
 
7
  "dropout": 0.0,
 
8
  "eos_token_id": 2,
 
9
  "hidden_act": "quick_gelu",
10
- "hidden_size": 32,
11
  "initializer_factor": 1.0,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 37,
 
 
14
  "layer_norm_eps": 1e-05,
 
15
  "max_position_embeddings": 77,
16
- "model_type": "clip_text_model",
17
- "num_attention_heads": 4,
18
- "num_hidden_layers": 5,
 
19
  "pad_token_id": 1,
 
 
20
  "torch_dtype": "float32",
21
- "transformers_version": "4.20.0.dev0",
 
22
  "vocab_size": 1000
23
  }
 
1
  {
2
+ "_name_or_path": "./text_encoder/",
3
  "architectures": [
4
+ "T5EncoderModel"
5
  ],
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 0,
8
+ "d_ff": 2048,
9
+ "d_kv": 64,
10
+ "d_model": 32,
11
+ "dense_act_fn": "relu",
12
  "dropout": 0.0,
13
+ "dropout_rate": 0.1,
14
  "eos_token_id": 2,
15
+ "feed_forward_proj": "relu",
16
  "hidden_act": "quick_gelu",
 
17
  "initializer_factor": 1.0,
18
  "initializer_range": 0.02,
19
  "intermediate_size": 37,
20
+ "is_encoder_decoder": true,
21
+ "is_gated_act": false,
22
  "layer_norm_eps": 1e-05,
23
+ "layer_norm_epsilon": 1e-06,
24
  "max_position_embeddings": 77,
25
+ "model_type": "t5",
26
+ "num_decoder_layers": 6,
27
+ "num_heads": 4,
28
+ "num_layers": 5,
29
  "pad_token_id": 1,
30
+ "relative_attention_max_distance": 128,
31
+ "relative_attention_num_buckets": 32,
32
  "torch_dtype": "float32",
33
+ "transformers_version": "4.28.0.dev0",
34
+ "use_cache": true,
35
  "vocab_size": 1000
36
  }
text_encoder/model-00001-of-00023.safetensors ADDED
Binary file (44 Bytes). View file
 
text_encoder/model-00002-of-00023.safetensors ADDED
Binary file (128 kB). View file
 
text_encoder/model-00003-of-00023.safetensors ADDED
Binary file (98.7 kB). View file
 
text_encoder/model-00004-of-00023.safetensors ADDED
Binary file (33.8 kB). View file
 
text_encoder/model-00005-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model-00006-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model-00007-of-00023.safetensors ADDED
Binary file (98.9 kB). View file
 
text_encoder/model-00008-of-00023.safetensors ADDED
Binary file (33.1 kB). View file
 
text_encoder/model-00009-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model-00010-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model-00011-of-00023.safetensors ADDED
Binary file (98.9 kB). View file
 
text_encoder/model-00012-of-00023.safetensors ADDED
Binary file (33.1 kB). View file
 
text_encoder/model-00013-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model-00014-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model-00015-of-00023.safetensors ADDED
Binary file (98.9 kB). View file
 
text_encoder/model-00016-of-00023.safetensors ADDED
Binary file (33.1 kB). View file
 
text_encoder/model-00017-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model-00018-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model-00019-of-00023.safetensors ADDED
Binary file (98.9 kB). View file
 
text_encoder/model-00020-of-00023.safetensors ADDED
Binary file (33.1 kB). View file
 
text_encoder/model-00021-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model-00022-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model-00023-of-00023.safetensors ADDED
Binary file (480 Bytes). View file
 
text_encoder/model.safetensors.index.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 3406720
4
+ },
5
+ "weight_map": {
6
+ "encoder.block.0.layer.0.SelfAttention.k.weight": "model-00003-of-00023.safetensors",
7
+ "encoder.block.0.layer.0.SelfAttention.o.weight": "model-00004-of-00023.safetensors",
8
+ "encoder.block.0.layer.0.SelfAttention.q.weight": "model-00003-of-00023.safetensors",
9
+ "encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight": "model-00004-of-00023.safetensors",
10
+ "encoder.block.0.layer.0.SelfAttention.v.weight": "model-00003-of-00023.safetensors",
11
+ "encoder.block.0.layer.0.layer_norm.weight": "model-00004-of-00023.safetensors",
12
+ "encoder.block.0.layer.1.DenseReluDense.wi.weight": "model-00005-of-00023.safetensors",
13
+ "encoder.block.0.layer.1.DenseReluDense.wo.weight": "model-00006-of-00023.safetensors",
14
+ "encoder.block.0.layer.1.layer_norm.weight": "model-00007-of-00023.safetensors",
15
+ "encoder.block.1.layer.0.SelfAttention.k.weight": "model-00007-of-00023.safetensors",
16
+ "encoder.block.1.layer.0.SelfAttention.o.weight": "model-00008-of-00023.safetensors",
17
+ "encoder.block.1.layer.0.SelfAttention.q.weight": "model-00007-of-00023.safetensors",
18
+ "encoder.block.1.layer.0.SelfAttention.v.weight": "model-00007-of-00023.safetensors",
19
+ "encoder.block.1.layer.0.layer_norm.weight": "model-00008-of-00023.safetensors",
20
+ "encoder.block.1.layer.1.DenseReluDense.wi.weight": "model-00009-of-00023.safetensors",
21
+ "encoder.block.1.layer.1.DenseReluDense.wo.weight": "model-00010-of-00023.safetensors",
22
+ "encoder.block.1.layer.1.layer_norm.weight": "model-00011-of-00023.safetensors",
23
+ "encoder.block.2.layer.0.SelfAttention.k.weight": "model-00011-of-00023.safetensors",
24
+ "encoder.block.2.layer.0.SelfAttention.o.weight": "model-00012-of-00023.safetensors",
25
+ "encoder.block.2.layer.0.SelfAttention.q.weight": "model-00011-of-00023.safetensors",
26
+ "encoder.block.2.layer.0.SelfAttention.v.weight": "model-00011-of-00023.safetensors",
27
+ "encoder.block.2.layer.0.layer_norm.weight": "model-00012-of-00023.safetensors",
28
+ "encoder.block.2.layer.1.DenseReluDense.wi.weight": "model-00013-of-00023.safetensors",
29
+ "encoder.block.2.layer.1.DenseReluDense.wo.weight": "model-00014-of-00023.safetensors",
30
+ "encoder.block.2.layer.1.layer_norm.weight": "model-00015-of-00023.safetensors",
31
+ "encoder.block.3.layer.0.SelfAttention.k.weight": "model-00015-of-00023.safetensors",
32
+ "encoder.block.3.layer.0.SelfAttention.o.weight": "model-00016-of-00023.safetensors",
33
+ "encoder.block.3.layer.0.SelfAttention.q.weight": "model-00015-of-00023.safetensors",
34
+ "encoder.block.3.layer.0.SelfAttention.v.weight": "model-00015-of-00023.safetensors",
35
+ "encoder.block.3.layer.0.layer_norm.weight": "model-00016-of-00023.safetensors",
36
+ "encoder.block.3.layer.1.DenseReluDense.wi.weight": "model-00017-of-00023.safetensors",
37
+ "encoder.block.3.layer.1.DenseReluDense.wo.weight": "model-00018-of-00023.safetensors",
38
+ "encoder.block.3.layer.1.layer_norm.weight": "model-00019-of-00023.safetensors",
39
+ "encoder.block.4.layer.0.SelfAttention.k.weight": "model-00019-of-00023.safetensors",
40
+ "encoder.block.4.layer.0.SelfAttention.o.weight": "model-00020-of-00023.safetensors",
41
+ "encoder.block.4.layer.0.SelfAttention.q.weight": "model-00019-of-00023.safetensors",
42
+ "encoder.block.4.layer.0.SelfAttention.v.weight": "model-00019-of-00023.safetensors",
43
+ "encoder.block.4.layer.0.layer_norm.weight": "model-00020-of-00023.safetensors",
44
+ "encoder.block.4.layer.1.DenseReluDense.wi.weight": "model-00021-of-00023.safetensors",
45
+ "encoder.block.4.layer.1.DenseReluDense.wo.weight": "model-00022-of-00023.safetensors",
46
+ "encoder.block.4.layer.1.layer_norm.weight": "model-00023-of-00023.safetensors",
47
+ "encoder.final_layer_norm.weight": "model-00023-of-00023.safetensors",
48
+ "shared.weight": "model-00002-of-00023.safetensors"
49
+ }
50
+ }
text_encoder/pytorch_model-00001-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be6a45202f9ac05db7838d5c383d550bd6a79229feba0107d529ed6f95039b6d
3
+ size 537
text_encoder/pytorch_model-00002-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12353b628d2acab6c5e77aba69e9ddd7a0a614c2f0dd43fa90eb89b143e9d16d
3
+ size 128938
text_encoder/pytorch_model-00003-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b23d5f59be191335b5e4e3fb003bc27b882e11a3c81bcebaf40cf0a17a72407
3
+ size 128938
text_encoder/pytorch_model-00004-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87a2787c0b40be163635a5337dbd759fbea1ee3583ff419b9f4a04a1be5423fd
3
+ size 99916
text_encoder/pytorch_model-00005-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0be2c2603dac2f86cb07ede81d28cb3c925ef9ac9ff02d02ca955972580b544b
3
+ size 35020
text_encoder/pytorch_model-00006-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cadfac22c0c28c3218ef4573ba2db0d31e67d02d1f1a212b38f6d920f1797f57
3
+ size 263082
text_encoder/pytorch_model-00007-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4a6e202bd6ee304226b6bae016ece925e30200a4ad105dea1dc098dda2e3e75
3
+ size 263082
text_encoder/pytorch_model-00008-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b561aada58bb98287a95ebb726e059c3e4fd566629e225904c60cfc568f690a0
3
+ size 100317
text_encoder/pytorch_model-00009-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36d727300dbb9536730f5be3de03c00c5bde907e1e77d44ff81c76973b56f192
3
+ size 34171
text_encoder/pytorch_model-00010-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:377054c3b9a9a29673d0116a08c4b8ac7c8f87eb58af3a4c07cc9f43183717ce
3
+ size 263082
text_encoder/pytorch_model-00011-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:206e02819496e0074f64240832608941fb3a51c68062a88f089e144fbf45e61e
3
+ size 263082
text_encoder/pytorch_model-00012-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16eb3b4c043fca2adc8f1afc8b606427ae5ff5ba4d9885e94bd7087e6b54330a
3
+ size 100317
text_encoder/pytorch_model-00013-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d4e37992e261a7d8a605ff12cc3949837009735b982d498883e483d4baf002c
3
+ size 34171
text_encoder/pytorch_model-00014-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6c171f1826f034ebfab439ee25edd8b0c334d5ea1aec6182c9234e697bdefdd
3
+ size 263082
text_encoder/pytorch_model-00015-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2746b8b8c718b9d2d0662cef2e83def762d60bcb7d67809276c29e861d8f6968
3
+ size 263082
text_encoder/pytorch_model-00016-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8440ccacbc457f37fbc21ebc3e7ca4b8339e3442cf68de83b28e0548f03bc12c
3
+ size 100317
text_encoder/pytorch_model-00017-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3d72914dcf6255f59541a388ba3fda4d2887b20c0e7358b8eaf14b9f74117b3
3
+ size 34171
text_encoder/pytorch_model-00018-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1793252508c4502732b82f92cdbb6786967d9384201b45674a5c898fe70f7f9
3
+ size 263082
text_encoder/pytorch_model-00019-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24d96339969907ad3b91f9d85c955ce80bac87526f8ce43d37ceebb5282ac366
3
+ size 263082
text_encoder/pytorch_model-00020-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:576ac52f24e8b22f456ea5f8d996af3082d4107d60d714b4615adb1496890122
3
+ size 100317
text_encoder/pytorch_model-00021-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e2b3e442c19cb73ebe91a5427936ab315191fc2732ed8494dd60508aaf2db48
3
+ size 34171
text_encoder/pytorch_model-00022-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd63880a05eab7c7fe71ae7ce80503027fa4a2f9b6c660d487d45141ff20ce97
3
+ size 263082
text_encoder/pytorch_model-00023-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e2a2c52e70d277497370b5e242c7e5f8bff65db7585bc71d753ddd0dca4031c
3
+ size 263082
text_encoder/pytorch_model-00024-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bbab091d0b7ceecb2edb47d1c30617a46ec5366d3e6147c96c27e2540b37dd9
3
+ size 1467
text_encoder/pytorch_model.bin.index.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 3534720
4
+ },
5
+ "weight_map": {
6
+ "encoder.block.0.layer.0.SelfAttention.k.weight": "pytorch_model-00004-of-00024.bin",
7
+ "encoder.block.0.layer.0.SelfAttention.o.weight": "pytorch_model-00005-of-00024.bin",
8
+ "encoder.block.0.layer.0.SelfAttention.q.weight": "pytorch_model-00004-of-00024.bin",
9
+ "encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight": "pytorch_model-00005-of-00024.bin",
10
+ "encoder.block.0.layer.0.SelfAttention.v.weight": "pytorch_model-00004-of-00024.bin",
11
+ "encoder.block.0.layer.0.layer_norm.weight": "pytorch_model-00005-of-00024.bin",
12
+ "encoder.block.0.layer.1.DenseReluDense.wi.weight": "pytorch_model-00006-of-00024.bin",
13
+ "encoder.block.0.layer.1.DenseReluDense.wo.weight": "pytorch_model-00007-of-00024.bin",
14
+ "encoder.block.0.layer.1.layer_norm.weight": "pytorch_model-00008-of-00024.bin",
15
+ "encoder.block.1.layer.0.SelfAttention.k.weight": "pytorch_model-00008-of-00024.bin",
16
+ "encoder.block.1.layer.0.SelfAttention.o.weight": "pytorch_model-00009-of-00024.bin",
17
+ "encoder.block.1.layer.0.SelfAttention.q.weight": "pytorch_model-00008-of-00024.bin",
18
+ "encoder.block.1.layer.0.SelfAttention.v.weight": "pytorch_model-00008-of-00024.bin",
19
+ "encoder.block.1.layer.0.layer_norm.weight": "pytorch_model-00009-of-00024.bin",
20
+ "encoder.block.1.layer.1.DenseReluDense.wi.weight": "pytorch_model-00010-of-00024.bin",
21
+ "encoder.block.1.layer.1.DenseReluDense.wo.weight": "pytorch_model-00011-of-00024.bin",
22
+ "encoder.block.1.layer.1.layer_norm.weight": "pytorch_model-00012-of-00024.bin",
23
+ "encoder.block.2.layer.0.SelfAttention.k.weight": "pytorch_model-00012-of-00024.bin",
24
+ "encoder.block.2.layer.0.SelfAttention.o.weight": "pytorch_model-00013-of-00024.bin",
25
+ "encoder.block.2.layer.0.SelfAttention.q.weight": "pytorch_model-00012-of-00024.bin",
26
+ "encoder.block.2.layer.0.SelfAttention.v.weight": "pytorch_model-00012-of-00024.bin",
27
+ "encoder.block.2.layer.0.layer_norm.weight": "pytorch_model-00013-of-00024.bin",
28
+ "encoder.block.2.layer.1.DenseReluDense.wi.weight": "pytorch_model-00014-of-00024.bin",
29
+ "encoder.block.2.layer.1.DenseReluDense.wo.weight": "pytorch_model-00015-of-00024.bin",
30
+ "encoder.block.2.layer.1.layer_norm.weight": "pytorch_model-00016-of-00024.bin",
31
+ "encoder.block.3.layer.0.SelfAttention.k.weight": "pytorch_model-00016-of-00024.bin",
32
+ "encoder.block.3.layer.0.SelfAttention.o.weight": "pytorch_model-00017-of-00024.bin",
33
+ "encoder.block.3.layer.0.SelfAttention.q.weight": "pytorch_model-00016-of-00024.bin",
34
+ "encoder.block.3.layer.0.SelfAttention.v.weight": "pytorch_model-00016-of-00024.bin",
35
+ "encoder.block.3.layer.0.layer_norm.weight": "pytorch_model-00017-of-00024.bin",
36
+ "encoder.block.3.layer.1.DenseReluDense.wi.weight": "pytorch_model-00018-of-00024.bin",
37
+ "encoder.block.3.layer.1.DenseReluDense.wo.weight": "pytorch_model-00019-of-00024.bin",
38
+ "encoder.block.3.layer.1.layer_norm.weight": "pytorch_model-00020-of-00024.bin",
39
+ "encoder.block.4.layer.0.SelfAttention.k.weight": "pytorch_model-00020-of-00024.bin",
40
+ "encoder.block.4.layer.0.SelfAttention.o.weight": "pytorch_model-00021-of-00024.bin",
41
+ "encoder.block.4.layer.0.SelfAttention.q.weight": "pytorch_model-00020-of-00024.bin",
42
+ "encoder.block.4.layer.0.SelfAttention.v.weight": "pytorch_model-00020-of-00024.bin",
43
+ "encoder.block.4.layer.0.layer_norm.weight": "pytorch_model-00021-of-00024.bin",
44
+ "encoder.block.4.layer.1.DenseReluDense.wi.weight": "pytorch_model-00022-of-00024.bin",
45
+ "encoder.block.4.layer.1.DenseReluDense.wo.weight": "pytorch_model-00023-of-00024.bin",
46
+ "encoder.block.4.layer.1.layer_norm.weight": "pytorch_model-00024-of-00024.bin",
47
+ "encoder.embed_tokens.weight": "pytorch_model-00003-of-00024.bin",
48
+ "encoder.final_layer_norm.weight": "pytorch_model-00024-of-00024.bin",
49
+ "shared.weight": "pytorch_model-00002-of-00024.bin"
50
+ }
51
+ }