patrickvonplaten commited on
Commit
7b73c42
·
1 Parent(s): c94722e
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. text_encoder/config.json +7 -18
  2. text_encoder/{model-00023-of-00023.safetensors → model-00001-of-00004.safetensors} +2 -2
  3. text_encoder/{flax_model.msgpack → model-00002-of-00004.safetensors} +2 -2
  4. text_encoder/{model-00004-of-00023.safetensors → model-00003-of-00004.safetensors} +2 -2
  5. text_encoder/{model-00007-of-00023.safetensors → model-00004-of-00004.safetensors} +2 -2
  6. text_encoder/model-00005-of-00023.safetensors +0 -3
  7. text_encoder/model-00006-of-00023.safetensors +0 -3
  8. text_encoder/model-00008-of-00023.safetensors +0 -3
  9. text_encoder/model-00009-of-00023.safetensors +0 -3
  10. text_encoder/model-00010-of-00023.safetensors +0 -3
  11. text_encoder/model-00011-of-00023.safetensors +0 -3
  12. text_encoder/model-00012-of-00023.safetensors +0 -3
  13. text_encoder/model-00013-of-00023.safetensors +0 -3
  14. text_encoder/model-00014-of-00023.safetensors +0 -3
  15. text_encoder/model-00015-of-00023.safetensors +0 -3
  16. text_encoder/model-00016-of-00023.safetensors +0 -3
  17. text_encoder/model-00017-of-00023.safetensors +0 -3
  18. text_encoder/model-00018-of-00023.safetensors +0 -3
  19. text_encoder/model-00019-of-00023.safetensors +0 -3
  20. text_encoder/model-00020-of-00023.safetensors +0 -3
  21. text_encoder/model-00021-of-00023.safetensors +0 -3
  22. text_encoder/model-00022-of-00023.safetensors +0 -3
  23. text_encoder/model.fp16-00002-of-00023.safetensors +0 -3
  24. text_encoder/model.fp16-00003-of-00023.safetensors +0 -3
  25. text_encoder/model.fp16-00004-of-00023.safetensors +0 -3
  26. text_encoder/model.fp16-00005-of-00023.safetensors +0 -3
  27. text_encoder/model.fp16-00006-of-00023.safetensors +0 -3
  28. text_encoder/model.fp16-00007-of-00023.safetensors +0 -3
  29. text_encoder/model.fp16-00008-of-00023.safetensors +0 -3
  30. text_encoder/model.fp16-00009-of-00023.safetensors +0 -3
  31. text_encoder/model.fp16-00010-of-00023.safetensors +0 -3
  32. text_encoder/model.fp16-00011-of-00023.safetensors +0 -3
  33. text_encoder/model.fp16-00012-of-00023.safetensors +0 -3
  34. text_encoder/model.fp16-00013-of-00023.safetensors +0 -3
  35. text_encoder/model.fp16-00014-of-00023.safetensors +0 -3
  36. text_encoder/model.fp16-00015-of-00023.safetensors +0 -3
  37. text_encoder/model.fp16-00016-of-00023.safetensors +0 -3
  38. text_encoder/model.fp16-00017-of-00023.safetensors +0 -3
  39. text_encoder/model.fp16-00018-of-00023.safetensors +0 -3
  40. text_encoder/model.fp16-00019-of-00023.safetensors +0 -3
  41. text_encoder/model.fp16-00020-of-00023.safetensors +0 -3
  42. text_encoder/model.fp16-00021-of-00023.safetensors +0 -3
  43. text_encoder/model.fp16-00022-of-00023.safetensors +0 -3
  44. text_encoder/model.fp16-00023-of-00023.safetensors +0 -3
  45. text_encoder/model.safetensors.index.fp16.json +86 -44
  46. text_encoder/model.safetensors.index.json +86 -44
  47. text_encoder/{model-00001-of-00023.safetensors → pytorch_model-00001-of-00004.bin} +2 -2
  48. text_encoder/pytorch_model-00001-of-00024.bin +0 -3
  49. text_encoder/{model-00002-of-00023.safetensors → pytorch_model-00002-of-00004.bin} +2 -2
  50. text_encoder/pytorch_model-00002-of-00024.bin +0 -3
text_encoder/config.json CHANGED
@@ -1,36 +1,25 @@
1
  {
2
- "_name_or_path": "./text_encoder/",
3
  "architectures": [
4
- "T5EncoderModel"
5
  ],
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 0,
8
- "d_ff": 2048,
9
- "d_kv": 64,
10
- "d_model": 32,
11
- "dense_act_fn": "relu",
12
  "dropout": 0.0,
13
- "dropout_rate": 0.1,
14
  "eos_token_id": 2,
15
- "feed_forward_proj": "relu",
16
  "hidden_act": "quick_gelu",
 
17
  "initializer_factor": 1.0,
18
  "initializer_range": 0.02,
19
  "intermediate_size": 37,
20
- "is_encoder_decoder": true,
21
- "is_gated_act": false,
22
  "layer_norm_eps": 1e-05,
23
- "layer_norm_epsilon": 1e-06,
24
  "max_position_embeddings": 77,
25
- "model_type": "t5",
26
- "num_decoder_layers": 6,
27
- "num_heads": 4,
28
- "num_layers": 5,
29
  "pad_token_id": 1,
30
- "relative_attention_max_distance": 128,
31
- "relative_attention_num_buckets": 32,
32
  "torch_dtype": "float32",
33
  "transformers_version": "4.28.0.dev0",
34
- "use_cache": true,
35
  "vocab_size": 1000
36
  }
 
1
  {
2
+ "_name_or_path": "hf-internal-testing/tiny-stable-diffusion-pipe",
3
  "architectures": [
4
+ "CLIPTextModel"
5
  ],
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 0,
 
 
 
 
8
  "dropout": 0.0,
 
9
  "eos_token_id": 2,
 
10
  "hidden_act": "quick_gelu",
11
+ "hidden_size": 32,
12
  "initializer_factor": 1.0,
13
  "initializer_range": 0.02,
14
  "intermediate_size": 37,
 
 
15
  "layer_norm_eps": 1e-05,
 
16
  "max_position_embeddings": 77,
17
+ "model_type": "clip_text_model",
18
+ "num_attention_heads": 4,
19
+ "num_hidden_layers": 5,
 
20
  "pad_token_id": 1,
21
+ "projection_dim": 512,
 
22
  "torch_dtype": "float32",
23
  "transformers_version": "4.28.0.dev0",
 
24
  "vocab_size": 1000
25
  }
text_encoder/{model-00023-of-00023.safetensors → model-00001-of-00004.safetensors} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a97ca9d1a2a8b0fbcb17be94da7e21f9f25154fd091104f4ef5f89b1db87dab0
3
- size 480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb2f042a629c39b489ee36e74f96042660beebc6d093bbf8c99a87eb8df4e35f
3
+ size 750
text_encoder/{flax_model.msgpack → model-00002-of-00004.safetensors} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e0f56336d7bb3ca2c416bb9d74d452c67d9443609084e712e59e57de96dac918
3
- size 276381
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f88574d77725507812c13b844c50d6cf4bdc6f15651c4cf9bee6715613ccf8b6
3
+ size 128150
text_encoder/{model-00004-of-00023.safetensors → model-00003-of-00004.safetensors} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:583a99c95b699404b1473b2e2c12ad46ad87d5310a4f77ae1a3f60f0a586bed3
3
- size 33792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac6bf016c5afc709b1f3aa7cb396f095fa4bc62adac8498cf32bd1428c74134d
3
+ size 81148
text_encoder/{model-00007-of-00023.safetensors → model-00004-of-00004.safetensors} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c03bf0daa4bc134e6a0c285b0ad6d0412624ce598542b0126d5f0c16b49b9d7c
3
- size 98902
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3506305c4ef6fd9a7d1725c8d9aefe19fc308519a035720c5148fdf87b91ddfe
3
+ size 73810
text_encoder/model-00005-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:15438f140dfec03a7e43b9ce9f66b5f401b3a5af7fab0e65aa3107bcad629064
3
- size 262302
 
 
 
 
text_encoder/model-00006-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f99879cc6ca587263f072fc5b037de022940fb91d650912a04e882c8e0f13960
3
- size 262302
 
 
 
 
text_encoder/model-00008-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a63c056401c2f1c9217ee5886eb2e99312a5233c9350e49ac81075029a8f7141
3
- size 33146
 
 
 
 
text_encoder/model-00009-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:97a71fe3337c9cddb5f5cffdc93ed136a0ba89cd654e534cde7a30bf4f7b2b61
3
- size 262302
 
 
 
 
text_encoder/model-00010-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:599e0ff1b20e729a57aae4dd55195cf4700b25cd3c24d1f005112ec76332ef13
3
- size 262302
 
 
 
 
text_encoder/model-00011-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:04fefc997c1d5a12e71d89ec2dfc524cc1efebc99c0642eb148a071123efe666
3
- size 98902
 
 
 
 
text_encoder/model-00012-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c564ac448ade2d503cf63ec1244f74a39e2e5b0dc8ae87087c1d09c20c810ca6
3
- size 33146
 
 
 
 
text_encoder/model-00013-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf63d0e5e03ff98c9dc216cf01e543380c87d4f14d62560c9e1aebfbe0287ec2
3
- size 262302
 
 
 
 
text_encoder/model-00014-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e34535a0d95dd900ebf5817b280bd6124d864e8369ef400d5b71616bb40fa7b1
3
- size 262302
 
 
 
 
text_encoder/model-00015-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:070790735754d24ab16f6072c524ab16d87a203b5612f46d15ba7090f727c619
3
- size 98902
 
 
 
 
text_encoder/model-00016-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:450b05051f4b6e815d100468bdf17dd7675686bb3a0b862ddb0f0b477d4cc9a0
3
- size 33146
 
 
 
 
text_encoder/model-00017-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:af95d89d6935ca9b27583cd3653d2adcb66b6b76aa62432454359d92dc27db19
3
- size 262302
 
 
 
 
text_encoder/model-00018-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d6d9257fce96a02496b3971618fafc1db6b1fbee67a30b81eea69dd867ed2d0b
3
- size 262302
 
 
 
 
text_encoder/model-00019-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8d65835dc23c5a5754cd8610152f269e347837121a86a83a3f59afdba5c2826
3
- size 98902
 
 
 
 
text_encoder/model-00020-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:89c0bf93b3b4067ff338409710e0bf4c0c20144e2bd34eb939500b864ac01856
3
- size 33146
 
 
 
 
text_encoder/model-00021-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6826f32c3dc6e5fde094702a28024e3f394c6ae683924b2806ca61fbe9f668be
3
- size 262302
 
 
 
 
text_encoder/model-00022-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6484b58a144d0e757a5a6d46402ba7866c4b51a275f339602023137ea94b3136
3
- size 262302
 
 
 
 
text_encoder/model.fp16-00002-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:947cbc4cf66a6b2d60f56c9719b1ce549d56be3cb727531e283493df75a76147
3
- size 128120
 
 
 
 
text_encoder/model.fp16-00003-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7b982f44afd453011c8e196182b465834dca10d580568663b0117b2510abc1d7
3
- size 98674
 
 
 
 
text_encoder/model.fp16-00004-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:583a99c95b699404b1473b2e2c12ad46ad87d5310a4f77ae1a3f60f0a586bed3
3
- size 33792
 
 
 
 
text_encoder/model.fp16-00005-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:15438f140dfec03a7e43b9ce9f66b5f401b3a5af7fab0e65aa3107bcad629064
3
- size 262302
 
 
 
 
text_encoder/model.fp16-00006-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f99879cc6ca587263f072fc5b037de022940fb91d650912a04e882c8e0f13960
3
- size 262302
 
 
 
 
text_encoder/model.fp16-00007-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c03bf0daa4bc134e6a0c285b0ad6d0412624ce598542b0126d5f0c16b49b9d7c
3
- size 98902
 
 
 
 
text_encoder/model.fp16-00008-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a63c056401c2f1c9217ee5886eb2e99312a5233c9350e49ac81075029a8f7141
3
- size 33146
 
 
 
 
text_encoder/model.fp16-00009-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:97a71fe3337c9cddb5f5cffdc93ed136a0ba89cd654e534cde7a30bf4f7b2b61
3
- size 262302
 
 
 
 
text_encoder/model.fp16-00010-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:599e0ff1b20e729a57aae4dd55195cf4700b25cd3c24d1f005112ec76332ef13
3
- size 262302
 
 
 
 
text_encoder/model.fp16-00011-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:04fefc997c1d5a12e71d89ec2dfc524cc1efebc99c0642eb148a071123efe666
3
- size 98902
 
 
 
 
text_encoder/model.fp16-00012-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c564ac448ade2d503cf63ec1244f74a39e2e5b0dc8ae87087c1d09c20c810ca6
3
- size 33146
 
 
 
 
text_encoder/model.fp16-00013-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf63d0e5e03ff98c9dc216cf01e543380c87d4f14d62560c9e1aebfbe0287ec2
3
- size 262302
 
 
 
 
text_encoder/model.fp16-00014-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e34535a0d95dd900ebf5817b280bd6124d864e8369ef400d5b71616bb40fa7b1
3
- size 262302
 
 
 
 
text_encoder/model.fp16-00015-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:070790735754d24ab16f6072c524ab16d87a203b5612f46d15ba7090f727c619
3
- size 98902
 
 
 
 
text_encoder/model.fp16-00016-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:450b05051f4b6e815d100468bdf17dd7675686bb3a0b862ddb0f0b477d4cc9a0
3
- size 33146
 
 
 
 
text_encoder/model.fp16-00017-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:af95d89d6935ca9b27583cd3653d2adcb66b6b76aa62432454359d92dc27db19
3
- size 262302
 
 
 
 
text_encoder/model.fp16-00018-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d6d9257fce96a02496b3971618fafc1db6b1fbee67a30b81eea69dd867ed2d0b
3
- size 262302
 
 
 
 
text_encoder/model.fp16-00019-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8d65835dc23c5a5754cd8610152f269e347837121a86a83a3f59afdba5c2826
3
- size 98902
 
 
 
 
text_encoder/model.fp16-00020-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:89c0bf93b3b4067ff338409710e0bf4c0c20144e2bd34eb939500b864ac01856
3
- size 33146
 
 
 
 
text_encoder/model.fp16-00021-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6826f32c3dc6e5fde094702a28024e3f394c6ae683924b2806ca61fbe9f668be
3
- size 262302
 
 
 
 
text_encoder/model.fp16-00022-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6484b58a144d0e757a5a6d46402ba7866c4b51a275f339602023137ea94b3136
3
- size 262302
 
 
 
 
text_encoder/model.fp16-00023-of-00023.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a97ca9d1a2a8b0fbcb17be94da7e21f9f25154fd091104f4ef5f89b1db87dab0
3
- size 480
 
 
 
 
text_encoder/model.safetensors.index.fp16.json CHANGED
@@ -1,50 +1,92 @@
1
  {
2
  "metadata": {
3
- "total_size": 3406720
4
  },
5
  "weight_map": {
6
- "encoder.block.0.layer.0.SelfAttention.k.weight": "model.fp16-00003-of-00023.safetensors",
7
- "encoder.block.0.layer.0.SelfAttention.o.weight": "model.fp16-00004-of-00023.safetensors",
8
- "encoder.block.0.layer.0.SelfAttention.q.weight": "model.fp16-00003-of-00023.safetensors",
9
- "encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight": "model.fp16-00004-of-00023.safetensors",
10
- "encoder.block.0.layer.0.SelfAttention.v.weight": "model.fp16-00003-of-00023.safetensors",
11
- "encoder.block.0.layer.0.layer_norm.weight": "model.fp16-00004-of-00023.safetensors",
12
- "encoder.block.0.layer.1.DenseReluDense.wi.weight": "model.fp16-00005-of-00023.safetensors",
13
- "encoder.block.0.layer.1.DenseReluDense.wo.weight": "model.fp16-00006-of-00023.safetensors",
14
- "encoder.block.0.layer.1.layer_norm.weight": "model.fp16-00007-of-00023.safetensors",
15
- "encoder.block.1.layer.0.SelfAttention.k.weight": "model.fp16-00007-of-00023.safetensors",
16
- "encoder.block.1.layer.0.SelfAttention.o.weight": "model.fp16-00008-of-00023.safetensors",
17
- "encoder.block.1.layer.0.SelfAttention.q.weight": "model.fp16-00007-of-00023.safetensors",
18
- "encoder.block.1.layer.0.SelfAttention.v.weight": "model.fp16-00007-of-00023.safetensors",
19
- "encoder.block.1.layer.0.layer_norm.weight": "model.fp16-00008-of-00023.safetensors",
20
- "encoder.block.1.layer.1.DenseReluDense.wi.weight": "model.fp16-00009-of-00023.safetensors",
21
- "encoder.block.1.layer.1.DenseReluDense.wo.weight": "model.fp16-00010-of-00023.safetensors",
22
- "encoder.block.1.layer.1.layer_norm.weight": "model.fp16-00011-of-00023.safetensors",
23
- "encoder.block.2.layer.0.SelfAttention.k.weight": "model.fp16-00011-of-00023.safetensors",
24
- "encoder.block.2.layer.0.SelfAttention.o.weight": "model.fp16-00012-of-00023.safetensors",
25
- "encoder.block.2.layer.0.SelfAttention.q.weight": "model.fp16-00011-of-00023.safetensors",
26
- "encoder.block.2.layer.0.SelfAttention.v.weight": "model.fp16-00011-of-00023.safetensors",
27
- "encoder.block.2.layer.0.layer_norm.weight": "model.fp16-00012-of-00023.safetensors",
28
- "encoder.block.2.layer.1.DenseReluDense.wi.weight": "model.fp16-00013-of-00023.safetensors",
29
- "encoder.block.2.layer.1.DenseReluDense.wo.weight": "model.fp16-00014-of-00023.safetensors",
30
- "encoder.block.2.layer.1.layer_norm.weight": "model.fp16-00015-of-00023.safetensors",
31
- "encoder.block.3.layer.0.SelfAttention.k.weight": "model.fp16-00015-of-00023.safetensors",
32
- "encoder.block.3.layer.0.SelfAttention.o.weight": "model.fp16-00016-of-00023.safetensors",
33
- "encoder.block.3.layer.0.SelfAttention.q.weight": "model.fp16-00015-of-00023.safetensors",
34
- "encoder.block.3.layer.0.SelfAttention.v.weight": "model.fp16-00015-of-00023.safetensors",
35
- "encoder.block.3.layer.0.layer_norm.weight": "model.fp16-00016-of-00023.safetensors",
36
- "encoder.block.3.layer.1.DenseReluDense.wi.weight": "model.fp16-00017-of-00023.safetensors",
37
- "encoder.block.3.layer.1.DenseReluDense.wo.weight": "model.fp16-00018-of-00023.safetensors",
38
- "encoder.block.3.layer.1.layer_norm.weight": "model.fp16-00019-of-00023.safetensors",
39
- "encoder.block.4.layer.0.SelfAttention.k.weight": "model.fp16-00019-of-00023.safetensors",
40
- "encoder.block.4.layer.0.SelfAttention.o.weight": "model.fp16-00020-of-00023.safetensors",
41
- "encoder.block.4.layer.0.SelfAttention.q.weight": "model.fp16-00019-of-00023.safetensors",
42
- "encoder.block.4.layer.0.SelfAttention.v.weight": "model.fp16-00019-of-00023.safetensors",
43
- "encoder.block.4.layer.0.layer_norm.weight": "model.fp16-00020-of-00023.safetensors",
44
- "encoder.block.4.layer.1.DenseReluDense.wi.weight": "model.fp16-00021-of-00023.safetensors",
45
- "encoder.block.4.layer.1.DenseReluDense.wo.weight": "model.fp16-00022-of-00023.safetensors",
46
- "encoder.block.4.layer.1.layer_norm.weight": "model.fp16-00023-of-00023.safetensors",
47
- "encoder.final_layer_norm.weight": "model.fp16-00023-of-00023.safetensors",
48
- "shared.weight": "model.fp16-00002-of-00023.safetensors"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
49
  }
50
  }
 
1
  {
2
  "metadata": {
3
+ "total_size": 274508
4
  },
5
  "weight_map": {
6
+ "text_model.embeddings.position_embedding.weight": "model.fp16-00003-of-00004.safetensors",
7
+ "text_model.embeddings.position_ids": "model.fp16-00001-of-00004.safetensors",
8
+ "text_model.embeddings.token_embedding.weight": "model.fp16-00002-of-00004.safetensors",
9
+ "text_model.encoder.layers.0.layer_norm1.bias": "model.fp16-00003-of-00004.safetensors",
10
+ "text_model.encoder.layers.0.layer_norm1.weight": "model.fp16-00003-of-00004.safetensors",
11
+ "text_model.encoder.layers.0.layer_norm2.bias": "model.fp16-00003-of-00004.safetensors",
12
+ "text_model.encoder.layers.0.layer_norm2.weight": "model.fp16-00003-of-00004.safetensors",
13
+ "text_model.encoder.layers.0.mlp.fc1.bias": "model.fp16-00003-of-00004.safetensors",
14
+ "text_model.encoder.layers.0.mlp.fc1.weight": "model.fp16-00003-of-00004.safetensors",
15
+ "text_model.encoder.layers.0.mlp.fc2.bias": "model.fp16-00003-of-00004.safetensors",
16
+ "text_model.encoder.layers.0.mlp.fc2.weight": "model.fp16-00003-of-00004.safetensors",
17
+ "text_model.encoder.layers.0.self_attn.k_proj.bias": "model.fp16-00003-of-00004.safetensors",
18
+ "text_model.encoder.layers.0.self_attn.k_proj.weight": "model.fp16-00003-of-00004.safetensors",
19
+ "text_model.encoder.layers.0.self_attn.out_proj.bias": "model.fp16-00003-of-00004.safetensors",
20
+ "text_model.encoder.layers.0.self_attn.out_proj.weight": "model.fp16-00003-of-00004.safetensors",
21
+ "text_model.encoder.layers.0.self_attn.q_proj.bias": "model.fp16-00003-of-00004.safetensors",
22
+ "text_model.encoder.layers.0.self_attn.q_proj.weight": "model.fp16-00003-of-00004.safetensors",
23
+ "text_model.encoder.layers.0.self_attn.v_proj.bias": "model.fp16-00003-of-00004.safetensors",
24
+ "text_model.encoder.layers.0.self_attn.v_proj.weight": "model.fp16-00003-of-00004.safetensors",
25
+ "text_model.encoder.layers.1.layer_norm1.bias": "model.fp16-00003-of-00004.safetensors",
26
+ "text_model.encoder.layers.1.layer_norm1.weight": "model.fp16-00003-of-00004.safetensors",
27
+ "text_model.encoder.layers.1.layer_norm2.bias": "model.fp16-00003-of-00004.safetensors",
28
+ "text_model.encoder.layers.1.layer_norm2.weight": "model.fp16-00003-of-00004.safetensors",
29
+ "text_model.encoder.layers.1.mlp.fc1.bias": "model.fp16-00003-of-00004.safetensors",
30
+ "text_model.encoder.layers.1.mlp.fc1.weight": "model.fp16-00003-of-00004.safetensors",
31
+ "text_model.encoder.layers.1.mlp.fc2.bias": "model.fp16-00003-of-00004.safetensors",
32
+ "text_model.encoder.layers.1.mlp.fc2.weight": "model.fp16-00003-of-00004.safetensors",
33
+ "text_model.encoder.layers.1.self_attn.k_proj.bias": "model.fp16-00003-of-00004.safetensors",
34
+ "text_model.encoder.layers.1.self_attn.k_proj.weight": "model.fp16-00003-of-00004.safetensors",
35
+ "text_model.encoder.layers.1.self_attn.out_proj.bias": "model.fp16-00003-of-00004.safetensors",
36
+ "text_model.encoder.layers.1.self_attn.out_proj.weight": "model.fp16-00003-of-00004.safetensors",
37
+ "text_model.encoder.layers.1.self_attn.q_proj.bias": "model.fp16-00003-of-00004.safetensors",
38
+ "text_model.encoder.layers.1.self_attn.q_proj.weight": "model.fp16-00003-of-00004.safetensors",
39
+ "text_model.encoder.layers.1.self_attn.v_proj.bias": "model.fp16-00003-of-00004.safetensors",
40
+ "text_model.encoder.layers.1.self_attn.v_proj.weight": "model.fp16-00003-of-00004.safetensors",
41
+ "text_model.encoder.layers.2.layer_norm1.bias": "model.fp16-00004-of-00004.safetensors",
42
+ "text_model.encoder.layers.2.layer_norm1.weight": "model.fp16-00004-of-00004.safetensors",
43
+ "text_model.encoder.layers.2.layer_norm2.bias": "model.fp16-00004-of-00004.safetensors",
44
+ "text_model.encoder.layers.2.layer_norm2.weight": "model.fp16-00004-of-00004.safetensors",
45
+ "text_model.encoder.layers.2.mlp.fc1.bias": "model.fp16-00004-of-00004.safetensors",
46
+ "text_model.encoder.layers.2.mlp.fc1.weight": "model.fp16-00004-of-00004.safetensors",
47
+ "text_model.encoder.layers.2.mlp.fc2.bias": "model.fp16-00004-of-00004.safetensors",
48
+ "text_model.encoder.layers.2.mlp.fc2.weight": "model.fp16-00004-of-00004.safetensors",
49
+ "text_model.encoder.layers.2.self_attn.k_proj.bias": "model.fp16-00003-of-00004.safetensors",
50
+ "text_model.encoder.layers.2.self_attn.k_proj.weight": "model.fp16-00003-of-00004.safetensors",
51
+ "text_model.encoder.layers.2.self_attn.out_proj.bias": "model.fp16-00004-of-00004.safetensors",
52
+ "text_model.encoder.layers.2.self_attn.out_proj.weight": "model.fp16-00004-of-00004.safetensors",
53
+ "text_model.encoder.layers.2.self_attn.q_proj.bias": "model.fp16-00003-of-00004.safetensors",
54
+ "text_model.encoder.layers.2.self_attn.q_proj.weight": "model.fp16-00003-of-00004.safetensors",
55
+ "text_model.encoder.layers.2.self_attn.v_proj.bias": "model.fp16-00003-of-00004.safetensors",
56
+ "text_model.encoder.layers.2.self_attn.v_proj.weight": "model.fp16-00003-of-00004.safetensors",
57
+ "text_model.encoder.layers.3.layer_norm1.bias": "model.fp16-00004-of-00004.safetensors",
58
+ "text_model.encoder.layers.3.layer_norm1.weight": "model.fp16-00004-of-00004.safetensors",
59
+ "text_model.encoder.layers.3.layer_norm2.bias": "model.fp16-00004-of-00004.safetensors",
60
+ "text_model.encoder.layers.3.layer_norm2.weight": "model.fp16-00004-of-00004.safetensors",
61
+ "text_model.encoder.layers.3.mlp.fc1.bias": "model.fp16-00004-of-00004.safetensors",
62
+ "text_model.encoder.layers.3.mlp.fc1.weight": "model.fp16-00004-of-00004.safetensors",
63
+ "text_model.encoder.layers.3.mlp.fc2.bias": "model.fp16-00004-of-00004.safetensors",
64
+ "text_model.encoder.layers.3.mlp.fc2.weight": "model.fp16-00004-of-00004.safetensors",
65
+ "text_model.encoder.layers.3.self_attn.k_proj.bias": "model.fp16-00004-of-00004.safetensors",
66
+ "text_model.encoder.layers.3.self_attn.k_proj.weight": "model.fp16-00004-of-00004.safetensors",
67
+ "text_model.encoder.layers.3.self_attn.out_proj.bias": "model.fp16-00004-of-00004.safetensors",
68
+ "text_model.encoder.layers.3.self_attn.out_proj.weight": "model.fp16-00004-of-00004.safetensors",
69
+ "text_model.encoder.layers.3.self_attn.q_proj.bias": "model.fp16-00004-of-00004.safetensors",
70
+ "text_model.encoder.layers.3.self_attn.q_proj.weight": "model.fp16-00004-of-00004.safetensors",
71
+ "text_model.encoder.layers.3.self_attn.v_proj.bias": "model.fp16-00004-of-00004.safetensors",
72
+ "text_model.encoder.layers.3.self_attn.v_proj.weight": "model.fp16-00004-of-00004.safetensors",
73
+ "text_model.encoder.layers.4.layer_norm1.bias": "model.fp16-00004-of-00004.safetensors",
74
+ "text_model.encoder.layers.4.layer_norm1.weight": "model.fp16-00004-of-00004.safetensors",
75
+ "text_model.encoder.layers.4.layer_norm2.bias": "model.fp16-00004-of-00004.safetensors",
76
+ "text_model.encoder.layers.4.layer_norm2.weight": "model.fp16-00004-of-00004.safetensors",
77
+ "text_model.encoder.layers.4.mlp.fc1.bias": "model.fp16-00004-of-00004.safetensors",
78
+ "text_model.encoder.layers.4.mlp.fc1.weight": "model.fp16-00004-of-00004.safetensors",
79
+ "text_model.encoder.layers.4.mlp.fc2.bias": "model.fp16-00004-of-00004.safetensors",
80
+ "text_model.encoder.layers.4.mlp.fc2.weight": "model.fp16-00004-of-00004.safetensors",
81
+ "text_model.encoder.layers.4.self_attn.k_proj.bias": "model.fp16-00004-of-00004.safetensors",
82
+ "text_model.encoder.layers.4.self_attn.k_proj.weight": "model.fp16-00004-of-00004.safetensors",
83
+ "text_model.encoder.layers.4.self_attn.out_proj.bias": "model.fp16-00004-of-00004.safetensors",
84
+ "text_model.encoder.layers.4.self_attn.out_proj.weight": "model.fp16-00004-of-00004.safetensors",
85
+ "text_model.encoder.layers.4.self_attn.q_proj.bias": "model.fp16-00004-of-00004.safetensors",
86
+ "text_model.encoder.layers.4.self_attn.q_proj.weight": "model.fp16-00004-of-00004.safetensors",
87
+ "text_model.encoder.layers.4.self_attn.v_proj.bias": "model.fp16-00004-of-00004.safetensors",
88
+ "text_model.encoder.layers.4.self_attn.v_proj.weight": "model.fp16-00004-of-00004.safetensors",
89
+ "text_model.final_layer_norm.bias": "model.fp16-00004-of-00004.safetensors",
90
+ "text_model.final_layer_norm.weight": "model.fp16-00004-of-00004.safetensors"
91
  }
92
  }
text_encoder/model.safetensors.index.json CHANGED
@@ -1,50 +1,92 @@
1
  {
2
  "metadata": {
3
- "total_size": 3406720
4
  },
5
  "weight_map": {
6
- "encoder.block.0.layer.0.SelfAttention.k.weight": "model-00003-of-00023.safetensors",
7
- "encoder.block.0.layer.0.SelfAttention.o.weight": "model-00004-of-00023.safetensors",
8
- "encoder.block.0.layer.0.SelfAttention.q.weight": "model-00003-of-00023.safetensors",
9
- "encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight": "model-00004-of-00023.safetensors",
10
- "encoder.block.0.layer.0.SelfAttention.v.weight": "model-00003-of-00023.safetensors",
11
- "encoder.block.0.layer.0.layer_norm.weight": "model-00004-of-00023.safetensors",
12
- "encoder.block.0.layer.1.DenseReluDense.wi.weight": "model-00005-of-00023.safetensors",
13
- "encoder.block.0.layer.1.DenseReluDense.wo.weight": "model-00006-of-00023.safetensors",
14
- "encoder.block.0.layer.1.layer_norm.weight": "model-00007-of-00023.safetensors",
15
- "encoder.block.1.layer.0.SelfAttention.k.weight": "model-00007-of-00023.safetensors",
16
- "encoder.block.1.layer.0.SelfAttention.o.weight": "model-00008-of-00023.safetensors",
17
- "encoder.block.1.layer.0.SelfAttention.q.weight": "model-00007-of-00023.safetensors",
18
- "encoder.block.1.layer.0.SelfAttention.v.weight": "model-00007-of-00023.safetensors",
19
- "encoder.block.1.layer.0.layer_norm.weight": "model-00008-of-00023.safetensors",
20
- "encoder.block.1.layer.1.DenseReluDense.wi.weight": "model-00009-of-00023.safetensors",
21
- "encoder.block.1.layer.1.DenseReluDense.wo.weight": "model-00010-of-00023.safetensors",
22
- "encoder.block.1.layer.1.layer_norm.weight": "model-00011-of-00023.safetensors",
23
- "encoder.block.2.layer.0.SelfAttention.k.weight": "model-00011-of-00023.safetensors",
24
- "encoder.block.2.layer.0.SelfAttention.o.weight": "model-00012-of-00023.safetensors",
25
- "encoder.block.2.layer.0.SelfAttention.q.weight": "model-00011-of-00023.safetensors",
26
- "encoder.block.2.layer.0.SelfAttention.v.weight": "model-00011-of-00023.safetensors",
27
- "encoder.block.2.layer.0.layer_norm.weight": "model-00012-of-00023.safetensors",
28
- "encoder.block.2.layer.1.DenseReluDense.wi.weight": "model-00013-of-00023.safetensors",
29
- "encoder.block.2.layer.1.DenseReluDense.wo.weight": "model-00014-of-00023.safetensors",
30
- "encoder.block.2.layer.1.layer_norm.weight": "model-00015-of-00023.safetensors",
31
- "encoder.block.3.layer.0.SelfAttention.k.weight": "model-00015-of-00023.safetensors",
32
- "encoder.block.3.layer.0.SelfAttention.o.weight": "model-00016-of-00023.safetensors",
33
- "encoder.block.3.layer.0.SelfAttention.q.weight": "model-00015-of-00023.safetensors",
34
- "encoder.block.3.layer.0.SelfAttention.v.weight": "model-00015-of-00023.safetensors",
35
- "encoder.block.3.layer.0.layer_norm.weight": "model-00016-of-00023.safetensors",
36
- "encoder.block.3.layer.1.DenseReluDense.wi.weight": "model-00017-of-00023.safetensors",
37
- "encoder.block.3.layer.1.DenseReluDense.wo.weight": "model-00018-of-00023.safetensors",
38
- "encoder.block.3.layer.1.layer_norm.weight": "model-00019-of-00023.safetensors",
39
- "encoder.block.4.layer.0.SelfAttention.k.weight": "model-00019-of-00023.safetensors",
40
- "encoder.block.4.layer.0.SelfAttention.o.weight": "model-00020-of-00023.safetensors",
41
- "encoder.block.4.layer.0.SelfAttention.q.weight": "model-00019-of-00023.safetensors",
42
- "encoder.block.4.layer.0.SelfAttention.v.weight": "model-00019-of-00023.safetensors",
43
- "encoder.block.4.layer.0.layer_norm.weight": "model-00020-of-00023.safetensors",
44
- "encoder.block.4.layer.1.DenseReluDense.wi.weight": "model-00021-of-00023.safetensors",
45
- "encoder.block.4.layer.1.DenseReluDense.wo.weight": "model-00022-of-00023.safetensors",
46
- "encoder.block.4.layer.1.layer_norm.weight": "model-00023-of-00023.safetensors",
47
- "encoder.final_layer_norm.weight": "model-00023-of-00023.safetensors",
48
- "shared.weight": "model-00002-of-00023.safetensors"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
49
  }
50
  }
 
1
  {
2
  "metadata": {
3
+ "total_size": 274508
4
  },
5
  "weight_map": {
6
+ "text_model.embeddings.position_embedding.weight": "model-00003-of-00004.safetensors",
7
+ "text_model.embeddings.position_ids": "model-00001-of-00004.safetensors",
8
+ "text_model.embeddings.token_embedding.weight": "model-00002-of-00004.safetensors",
9
+ "text_model.encoder.layers.0.layer_norm1.bias": "model-00003-of-00004.safetensors",
10
+ "text_model.encoder.layers.0.layer_norm1.weight": "model-00003-of-00004.safetensors",
11
+ "text_model.encoder.layers.0.layer_norm2.bias": "model-00003-of-00004.safetensors",
12
+ "text_model.encoder.layers.0.layer_norm2.weight": "model-00003-of-00004.safetensors",
13
+ "text_model.encoder.layers.0.mlp.fc1.bias": "model-00003-of-00004.safetensors",
14
+ "text_model.encoder.layers.0.mlp.fc1.weight": "model-00003-of-00004.safetensors",
15
+ "text_model.encoder.layers.0.mlp.fc2.bias": "model-00003-of-00004.safetensors",
16
+ "text_model.encoder.layers.0.mlp.fc2.weight": "model-00003-of-00004.safetensors",
17
+ "text_model.encoder.layers.0.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
18
+ "text_model.encoder.layers.0.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
19
+ "text_model.encoder.layers.0.self_attn.out_proj.bias": "model-00003-of-00004.safetensors",
20
+ "text_model.encoder.layers.0.self_attn.out_proj.weight": "model-00003-of-00004.safetensors",
21
+ "text_model.encoder.layers.0.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
22
+ "text_model.encoder.layers.0.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
23
+ "text_model.encoder.layers.0.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
24
+ "text_model.encoder.layers.0.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
25
+ "text_model.encoder.layers.1.layer_norm1.bias": "model-00003-of-00004.safetensors",
26
+ "text_model.encoder.layers.1.layer_norm1.weight": "model-00003-of-00004.safetensors",
27
+ "text_model.encoder.layers.1.layer_norm2.bias": "model-00003-of-00004.safetensors",
28
+ "text_model.encoder.layers.1.layer_norm2.weight": "model-00003-of-00004.safetensors",
29
+ "text_model.encoder.layers.1.mlp.fc1.bias": "model-00003-of-00004.safetensors",
30
+ "text_model.encoder.layers.1.mlp.fc1.weight": "model-00003-of-00004.safetensors",
31
+ "text_model.encoder.layers.1.mlp.fc2.bias": "model-00003-of-00004.safetensors",
32
+ "text_model.encoder.layers.1.mlp.fc2.weight": "model-00003-of-00004.safetensors",
33
+ "text_model.encoder.layers.1.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
34
+ "text_model.encoder.layers.1.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
35
+ "text_model.encoder.layers.1.self_attn.out_proj.bias": "model-00003-of-00004.safetensors",
36
+ "text_model.encoder.layers.1.self_attn.out_proj.weight": "model-00003-of-00004.safetensors",
37
+ "text_model.encoder.layers.1.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
38
+ "text_model.encoder.layers.1.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
39
+ "text_model.encoder.layers.1.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
40
+ "text_model.encoder.layers.1.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
41
+ "text_model.encoder.layers.2.layer_norm1.bias": "model-00004-of-00004.safetensors",
42
+ "text_model.encoder.layers.2.layer_norm1.weight": "model-00004-of-00004.safetensors",
43
+ "text_model.encoder.layers.2.layer_norm2.bias": "model-00004-of-00004.safetensors",
44
+ "text_model.encoder.layers.2.layer_norm2.weight": "model-00004-of-00004.safetensors",
45
+ "text_model.encoder.layers.2.mlp.fc1.bias": "model-00004-of-00004.safetensors",
46
+ "text_model.encoder.layers.2.mlp.fc1.weight": "model-00004-of-00004.safetensors",
47
+ "text_model.encoder.layers.2.mlp.fc2.bias": "model-00004-of-00004.safetensors",
48
+ "text_model.encoder.layers.2.mlp.fc2.weight": "model-00004-of-00004.safetensors",
49
+ "text_model.encoder.layers.2.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
50
+ "text_model.encoder.layers.2.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
51
+ "text_model.encoder.layers.2.self_attn.out_proj.bias": "model-00004-of-00004.safetensors",
52
+ "text_model.encoder.layers.2.self_attn.out_proj.weight": "model-00004-of-00004.safetensors",
53
+ "text_model.encoder.layers.2.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
54
+ "text_model.encoder.layers.2.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
55
+ "text_model.encoder.layers.2.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
56
+ "text_model.encoder.layers.2.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
57
+ "text_model.encoder.layers.3.layer_norm1.bias": "model-00004-of-00004.safetensors",
58
+ "text_model.encoder.layers.3.layer_norm1.weight": "model-00004-of-00004.safetensors",
59
+ "text_model.encoder.layers.3.layer_norm2.bias": "model-00004-of-00004.safetensors",
60
+ "text_model.encoder.layers.3.layer_norm2.weight": "model-00004-of-00004.safetensors",
61
+ "text_model.encoder.layers.3.mlp.fc1.bias": "model-00004-of-00004.safetensors",
62
+ "text_model.encoder.layers.3.mlp.fc1.weight": "model-00004-of-00004.safetensors",
63
+ "text_model.encoder.layers.3.mlp.fc2.bias": "model-00004-of-00004.safetensors",
64
+ "text_model.encoder.layers.3.mlp.fc2.weight": "model-00004-of-00004.safetensors",
65
+ "text_model.encoder.layers.3.self_attn.k_proj.bias": "model-00004-of-00004.safetensors",
66
+ "text_model.encoder.layers.3.self_attn.k_proj.weight": "model-00004-of-00004.safetensors",
67
+ "text_model.encoder.layers.3.self_attn.out_proj.bias": "model-00004-of-00004.safetensors",
68
+ "text_model.encoder.layers.3.self_attn.out_proj.weight": "model-00004-of-00004.safetensors",
69
+ "text_model.encoder.layers.3.self_attn.q_proj.bias": "model-00004-of-00004.safetensors",
70
+ "text_model.encoder.layers.3.self_attn.q_proj.weight": "model-00004-of-00004.safetensors",
71
+ "text_model.encoder.layers.3.self_attn.v_proj.bias": "model-00004-of-00004.safetensors",
72
+ "text_model.encoder.layers.3.self_attn.v_proj.weight": "model-00004-of-00004.safetensors",
73
+ "text_model.encoder.layers.4.layer_norm1.bias": "model-00004-of-00004.safetensors",
74
+ "text_model.encoder.layers.4.layer_norm1.weight": "model-00004-of-00004.safetensors",
75
+ "text_model.encoder.layers.4.layer_norm2.bias": "model-00004-of-00004.safetensors",
76
+ "text_model.encoder.layers.4.layer_norm2.weight": "model-00004-of-00004.safetensors",
77
+ "text_model.encoder.layers.4.mlp.fc1.bias": "model-00004-of-00004.safetensors",
78
+ "text_model.encoder.layers.4.mlp.fc1.weight": "model-00004-of-00004.safetensors",
79
+ "text_model.encoder.layers.4.mlp.fc2.bias": "model-00004-of-00004.safetensors",
80
+ "text_model.encoder.layers.4.mlp.fc2.weight": "model-00004-of-00004.safetensors",
81
+ "text_model.encoder.layers.4.self_attn.k_proj.bias": "model-00004-of-00004.safetensors",
82
+ "text_model.encoder.layers.4.self_attn.k_proj.weight": "model-00004-of-00004.safetensors",
83
+ "text_model.encoder.layers.4.self_attn.out_proj.bias": "model-00004-of-00004.safetensors",
84
+ "text_model.encoder.layers.4.self_attn.out_proj.weight": "model-00004-of-00004.safetensors",
85
+ "text_model.encoder.layers.4.self_attn.q_proj.bias": "model-00004-of-00004.safetensors",
86
+ "text_model.encoder.layers.4.self_attn.q_proj.weight": "model-00004-of-00004.safetensors",
87
+ "text_model.encoder.layers.4.self_attn.v_proj.bias": "model-00004-of-00004.safetensors",
88
+ "text_model.encoder.layers.4.self_attn.v_proj.weight": "model-00004-of-00004.safetensors",
89
+ "text_model.final_layer_norm.bias": "model-00004-of-00004.safetensors",
90
+ "text_model.final_layer_norm.weight": "model-00004-of-00004.safetensors"
91
  }
92
  }
text_encoder/{model-00001-of-00023.safetensors → pytorch_model-00001-of-00004.bin} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:12af4f14c56f2977175469f763f0689c3032f02d31eb5136317659088e6e6d93
3
- size 44
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ca8004c44cd75ca1f0960935d75ade355702b762df72d48e4ac30db26f4f8d9
3
+ size 1387
text_encoder/pytorch_model-00001-of-00024.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:be6a45202f9ac05db7838d5c383d550bd6a79229feba0107d529ed6f95039b6d
3
- size 537
 
 
 
 
text_encoder/{model-00002-of-00023.safetensors → pytorch_model-00002-of-00004.bin} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:947cbc4cf66a6b2d60f56c9719b1ce549d56be3cb727531e283493df75a76147
3
- size 128120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d71a6eba65120451df6a1c2cafe5daa7b3474cb9f7eef172a49320f3b17c649
3
+ size 128811
text_encoder/pytorch_model-00002-of-00024.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:12353b628d2acab6c5e77aba69e9ddd7a0a614c2f0dd43fa90eb89b143e9d16d
3
- size 128938