patrickvonplaten commited on
Commit
eab6a16
·
1 Parent(s): ab890c9
Files changed (49) hide show
  1. text_encoder/model.fp16-00001-of-00023.safetensors +0 -0
  2. text_encoder/model.fp16-00002-of-00023.safetensors +0 -0
  3. text_encoder/model.fp16-00003-of-00023.safetensors +0 -0
  4. text_encoder/model.fp16-00004-of-00023.safetensors +0 -0
  5. text_encoder/model.fp16-00005-of-00023.safetensors +0 -0
  6. text_encoder/model.fp16-00006-of-00023.safetensors +0 -0
  7. text_encoder/model.fp16-00007-of-00023.safetensors +0 -0
  8. text_encoder/model.fp16-00008-of-00023.safetensors +0 -0
  9. text_encoder/model.fp16-00009-of-00023.safetensors +0 -0
  10. text_encoder/model.fp16-00010-of-00023.safetensors +0 -0
  11. text_encoder/model.fp16-00011-of-00023.safetensors +0 -0
  12. text_encoder/model.fp16-00012-of-00023.safetensors +0 -0
  13. text_encoder/model.fp16-00013-of-00023.safetensors +0 -0
  14. text_encoder/model.fp16-00014-of-00023.safetensors +0 -0
  15. text_encoder/model.fp16-00015-of-00023.safetensors +0 -0
  16. text_encoder/model.fp16-00016-of-00023.safetensors +0 -0
  17. text_encoder/model.fp16-00017-of-00023.safetensors +0 -0
  18. text_encoder/model.fp16-00018-of-00023.safetensors +0 -0
  19. text_encoder/model.fp16-00019-of-00023.safetensors +0 -0
  20. text_encoder/model.fp16-00020-of-00023.safetensors +0 -0
  21. text_encoder/model.fp16-00021-of-00023.safetensors +0 -0
  22. text_encoder/model.fp16-00022-of-00023.safetensors +0 -0
  23. text_encoder/model.fp16-00023-of-00023.safetensors +0 -0
  24. text_encoder/model.safetensors.index.fp16.json +50 -0
  25. text_encoder/pytorch_model.bin.index.fp16.json +51 -0
  26. text_encoder/pytorch_model.fp16-00001-of-00024.bin +3 -0
  27. text_encoder/pytorch_model.fp16-00002-of-00024.bin +3 -0
  28. text_encoder/pytorch_model.fp16-00003-of-00024.bin +3 -0
  29. text_encoder/pytorch_model.fp16-00004-of-00024.bin +3 -0
  30. text_encoder/pytorch_model.fp16-00005-of-00024.bin +3 -0
  31. text_encoder/pytorch_model.fp16-00006-of-00024.bin +3 -0
  32. text_encoder/pytorch_model.fp16-00007-of-00024.bin +3 -0
  33. text_encoder/pytorch_model.fp16-00008-of-00024.bin +3 -0
  34. text_encoder/pytorch_model.fp16-00009-of-00024.bin +3 -0
  35. text_encoder/pytorch_model.fp16-00010-of-00024.bin +3 -0
  36. text_encoder/pytorch_model.fp16-00011-of-00024.bin +3 -0
  37. text_encoder/pytorch_model.fp16-00012-of-00024.bin +3 -0
  38. text_encoder/pytorch_model.fp16-00013-of-00024.bin +3 -0
  39. text_encoder/pytorch_model.fp16-00014-of-00024.bin +3 -0
  40. text_encoder/pytorch_model.fp16-00015-of-00024.bin +3 -0
  41. text_encoder/pytorch_model.fp16-00016-of-00024.bin +3 -0
  42. text_encoder/pytorch_model.fp16-00017-of-00024.bin +3 -0
  43. text_encoder/pytorch_model.fp16-00018-of-00024.bin +3 -0
  44. text_encoder/pytorch_model.fp16-00019-of-00024.bin +3 -0
  45. text_encoder/pytorch_model.fp16-00020-of-00024.bin +3 -0
  46. text_encoder/pytorch_model.fp16-00021-of-00024.bin +3 -0
  47. text_encoder/pytorch_model.fp16-00022-of-00024.bin +3 -0
  48. text_encoder/pytorch_model.fp16-00023-of-00024.bin +3 -0
  49. text_encoder/pytorch_model.fp16-00024-of-00024.bin +3 -0
text_encoder/model.fp16-00001-of-00023.safetensors ADDED
Binary file (44 Bytes). View file
 
text_encoder/model.fp16-00002-of-00023.safetensors ADDED
Binary file (128 kB). View file
 
text_encoder/model.fp16-00003-of-00023.safetensors ADDED
Binary file (98.7 kB). View file
 
text_encoder/model.fp16-00004-of-00023.safetensors ADDED
Binary file (33.8 kB). View file
 
text_encoder/model.fp16-00005-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model.fp16-00006-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model.fp16-00007-of-00023.safetensors ADDED
Binary file (98.9 kB). View file
 
text_encoder/model.fp16-00008-of-00023.safetensors ADDED
Binary file (33.1 kB). View file
 
text_encoder/model.fp16-00009-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model.fp16-00010-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model.fp16-00011-of-00023.safetensors ADDED
Binary file (98.9 kB). View file
 
text_encoder/model.fp16-00012-of-00023.safetensors ADDED
Binary file (33.1 kB). View file
 
text_encoder/model.fp16-00013-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model.fp16-00014-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model.fp16-00015-of-00023.safetensors ADDED
Binary file (98.9 kB). View file
 
text_encoder/model.fp16-00016-of-00023.safetensors ADDED
Binary file (33.1 kB). View file
 
text_encoder/model.fp16-00017-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model.fp16-00018-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model.fp16-00019-of-00023.safetensors ADDED
Binary file (98.9 kB). View file
 
text_encoder/model.fp16-00020-of-00023.safetensors ADDED
Binary file (33.1 kB). View file
 
text_encoder/model.fp16-00021-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model.fp16-00022-of-00023.safetensors ADDED
Binary file (262 kB). View file
 
text_encoder/model.fp16-00023-of-00023.safetensors ADDED
Binary file (480 Bytes). View file
 
text_encoder/model.safetensors.index.fp16.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 3406720
4
+ },
5
+ "weight_map": {
6
+ "encoder.block.0.layer.0.SelfAttention.k.weight": "model.fp16-00003-of-00023.safetensors",
7
+ "encoder.block.0.layer.0.SelfAttention.o.weight": "model.fp16-00004-of-00023.safetensors",
8
+ "encoder.block.0.layer.0.SelfAttention.q.weight": "model.fp16-00003-of-00023.safetensors",
9
+ "encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight": "model.fp16-00004-of-00023.safetensors",
10
+ "encoder.block.0.layer.0.SelfAttention.v.weight": "model.fp16-00003-of-00023.safetensors",
11
+ "encoder.block.0.layer.0.layer_norm.weight": "model.fp16-00004-of-00023.safetensors",
12
+ "encoder.block.0.layer.1.DenseReluDense.wi.weight": "model.fp16-00005-of-00023.safetensors",
13
+ "encoder.block.0.layer.1.DenseReluDense.wo.weight": "model.fp16-00006-of-00023.safetensors",
14
+ "encoder.block.0.layer.1.layer_norm.weight": "model.fp16-00007-of-00023.safetensors",
15
+ "encoder.block.1.layer.0.SelfAttention.k.weight": "model.fp16-00007-of-00023.safetensors",
16
+ "encoder.block.1.layer.0.SelfAttention.o.weight": "model.fp16-00008-of-00023.safetensors",
17
+ "encoder.block.1.layer.0.SelfAttention.q.weight": "model.fp16-00007-of-00023.safetensors",
18
+ "encoder.block.1.layer.0.SelfAttention.v.weight": "model.fp16-00007-of-00023.safetensors",
19
+ "encoder.block.1.layer.0.layer_norm.weight": "model.fp16-00008-of-00023.safetensors",
20
+ "encoder.block.1.layer.1.DenseReluDense.wi.weight": "model.fp16-00009-of-00023.safetensors",
21
+ "encoder.block.1.layer.1.DenseReluDense.wo.weight": "model.fp16-00010-of-00023.safetensors",
22
+ "encoder.block.1.layer.1.layer_norm.weight": "model.fp16-00011-of-00023.safetensors",
23
+ "encoder.block.2.layer.0.SelfAttention.k.weight": "model.fp16-00011-of-00023.safetensors",
24
+ "encoder.block.2.layer.0.SelfAttention.o.weight": "model.fp16-00012-of-00023.safetensors",
25
+ "encoder.block.2.layer.0.SelfAttention.q.weight": "model.fp16-00011-of-00023.safetensors",
26
+ "encoder.block.2.layer.0.SelfAttention.v.weight": "model.fp16-00011-of-00023.safetensors",
27
+ "encoder.block.2.layer.0.layer_norm.weight": "model.fp16-00012-of-00023.safetensors",
28
+ "encoder.block.2.layer.1.DenseReluDense.wi.weight": "model.fp16-00013-of-00023.safetensors",
29
+ "encoder.block.2.layer.1.DenseReluDense.wo.weight": "model.fp16-00014-of-00023.safetensors",
30
+ "encoder.block.2.layer.1.layer_norm.weight": "model.fp16-00015-of-00023.safetensors",
31
+ "encoder.block.3.layer.0.SelfAttention.k.weight": "model.fp16-00015-of-00023.safetensors",
32
+ "encoder.block.3.layer.0.SelfAttention.o.weight": "model.fp16-00016-of-00023.safetensors",
33
+ "encoder.block.3.layer.0.SelfAttention.q.weight": "model.fp16-00015-of-00023.safetensors",
34
+ "encoder.block.3.layer.0.SelfAttention.v.weight": "model.fp16-00015-of-00023.safetensors",
35
+ "encoder.block.3.layer.0.layer_norm.weight": "model.fp16-00016-of-00023.safetensors",
36
+ "encoder.block.3.layer.1.DenseReluDense.wi.weight": "model.fp16-00017-of-00023.safetensors",
37
+ "encoder.block.3.layer.1.DenseReluDense.wo.weight": "model.fp16-00018-of-00023.safetensors",
38
+ "encoder.block.3.layer.1.layer_norm.weight": "model.fp16-00019-of-00023.safetensors",
39
+ "encoder.block.4.layer.0.SelfAttention.k.weight": "model.fp16-00019-of-00023.safetensors",
40
+ "encoder.block.4.layer.0.SelfAttention.o.weight": "model.fp16-00020-of-00023.safetensors",
41
+ "encoder.block.4.layer.0.SelfAttention.q.weight": "model.fp16-00019-of-00023.safetensors",
42
+ "encoder.block.4.layer.0.SelfAttention.v.weight": "model.fp16-00019-of-00023.safetensors",
43
+ "encoder.block.4.layer.0.layer_norm.weight": "model.fp16-00020-of-00023.safetensors",
44
+ "encoder.block.4.layer.1.DenseReluDense.wi.weight": "model.fp16-00021-of-00023.safetensors",
45
+ "encoder.block.4.layer.1.DenseReluDense.wo.weight": "model.fp16-00022-of-00023.safetensors",
46
+ "encoder.block.4.layer.1.layer_norm.weight": "model.fp16-00023-of-00023.safetensors",
47
+ "encoder.final_layer_norm.weight": "model.fp16-00023-of-00023.safetensors",
48
+ "shared.weight": "model.fp16-00002-of-00023.safetensors"
49
+ }
50
+ }
text_encoder/pytorch_model.bin.index.fp16.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 3534720
4
+ },
5
+ "weight_map": {
6
+ "encoder.block.0.layer.0.SelfAttention.k.weight": "pytorch_model.fp16-00004-of-00024.bin",
7
+ "encoder.block.0.layer.0.SelfAttention.o.weight": "pytorch_model.fp16-00005-of-00024.bin",
8
+ "encoder.block.0.layer.0.SelfAttention.q.weight": "pytorch_model.fp16-00004-of-00024.bin",
9
+ "encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight": "pytorch_model.fp16-00005-of-00024.bin",
10
+ "encoder.block.0.layer.0.SelfAttention.v.weight": "pytorch_model.fp16-00004-of-00024.bin",
11
+ "encoder.block.0.layer.0.layer_norm.weight": "pytorch_model.fp16-00005-of-00024.bin",
12
+ "encoder.block.0.layer.1.DenseReluDense.wi.weight": "pytorch_model.fp16-00006-of-00024.bin",
13
+ "encoder.block.0.layer.1.DenseReluDense.wo.weight": "pytorch_model.fp16-00007-of-00024.bin",
14
+ "encoder.block.0.layer.1.layer_norm.weight": "pytorch_model.fp16-00008-of-00024.bin",
15
+ "encoder.block.1.layer.0.SelfAttention.k.weight": "pytorch_model.fp16-00008-of-00024.bin",
16
+ "encoder.block.1.layer.0.SelfAttention.o.weight": "pytorch_model.fp16-00009-of-00024.bin",
17
+ "encoder.block.1.layer.0.SelfAttention.q.weight": "pytorch_model.fp16-00008-of-00024.bin",
18
+ "encoder.block.1.layer.0.SelfAttention.v.weight": "pytorch_model.fp16-00008-of-00024.bin",
19
+ "encoder.block.1.layer.0.layer_norm.weight": "pytorch_model.fp16-00009-of-00024.bin",
20
+ "encoder.block.1.layer.1.DenseReluDense.wi.weight": "pytorch_model.fp16-00010-of-00024.bin",
21
+ "encoder.block.1.layer.1.DenseReluDense.wo.weight": "pytorch_model.fp16-00011-of-00024.bin",
22
+ "encoder.block.1.layer.1.layer_norm.weight": "pytorch_model.fp16-00012-of-00024.bin",
23
+ "encoder.block.2.layer.0.SelfAttention.k.weight": "pytorch_model.fp16-00012-of-00024.bin",
24
+ "encoder.block.2.layer.0.SelfAttention.o.weight": "pytorch_model.fp16-00013-of-00024.bin",
25
+ "encoder.block.2.layer.0.SelfAttention.q.weight": "pytorch_model.fp16-00012-of-00024.bin",
26
+ "encoder.block.2.layer.0.SelfAttention.v.weight": "pytorch_model.fp16-00012-of-00024.bin",
27
+ "encoder.block.2.layer.0.layer_norm.weight": "pytorch_model.fp16-00013-of-00024.bin",
28
+ "encoder.block.2.layer.1.DenseReluDense.wi.weight": "pytorch_model.fp16-00014-of-00024.bin",
29
+ "encoder.block.2.layer.1.DenseReluDense.wo.weight": "pytorch_model.fp16-00015-of-00024.bin",
30
+ "encoder.block.2.layer.1.layer_norm.weight": "pytorch_model.fp16-00016-of-00024.bin",
31
+ "encoder.block.3.layer.0.SelfAttention.k.weight": "pytorch_model.fp16-00016-of-00024.bin",
32
+ "encoder.block.3.layer.0.SelfAttention.o.weight": "pytorch_model.fp16-00017-of-00024.bin",
33
+ "encoder.block.3.layer.0.SelfAttention.q.weight": "pytorch_model.fp16-00016-of-00024.bin",
34
+ "encoder.block.3.layer.0.SelfAttention.v.weight": "pytorch_model.fp16-00016-of-00024.bin",
35
+ "encoder.block.3.layer.0.layer_norm.weight": "pytorch_model.fp16-00017-of-00024.bin",
36
+ "encoder.block.3.layer.1.DenseReluDense.wi.weight": "pytorch_model.fp16-00018-of-00024.bin",
37
+ "encoder.block.3.layer.1.DenseReluDense.wo.weight": "pytorch_model.fp16-00019-of-00024.bin",
38
+ "encoder.block.3.layer.1.layer_norm.weight": "pytorch_model.fp16-00020-of-00024.bin",
39
+ "encoder.block.4.layer.0.SelfAttention.k.weight": "pytorch_model.fp16-00020-of-00024.bin",
40
+ "encoder.block.4.layer.0.SelfAttention.o.weight": "pytorch_model.fp16-00021-of-00024.bin",
41
+ "encoder.block.4.layer.0.SelfAttention.q.weight": "pytorch_model.fp16-00020-of-00024.bin",
42
+ "encoder.block.4.layer.0.SelfAttention.v.weight": "pytorch_model.fp16-00020-of-00024.bin",
43
+ "encoder.block.4.layer.0.layer_norm.weight": "pytorch_model.fp16-00021-of-00024.bin",
44
+ "encoder.block.4.layer.1.DenseReluDense.wi.weight": "pytorch_model.fp16-00022-of-00024.bin",
45
+ "encoder.block.4.layer.1.DenseReluDense.wo.weight": "pytorch_model.fp16-00023-of-00024.bin",
46
+ "encoder.block.4.layer.1.layer_norm.weight": "pytorch_model.fp16-00024-of-00024.bin",
47
+ "encoder.embed_tokens.weight": "pytorch_model.fp16-00003-of-00024.bin",
48
+ "encoder.final_layer_norm.weight": "pytorch_model.fp16-00024-of-00024.bin",
49
+ "shared.weight": "pytorch_model.fp16-00002-of-00024.bin"
50
+ }
51
+ }
text_encoder/pytorch_model.fp16-00001-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61551b689644c71134776a15559f62162328ab0d7403d64599f02cfd3c136e40
3
+ size 547
text_encoder/pytorch_model.fp16-00002-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca920c65372b99404f815d69f3c4a6c369e58968b45f1a1c53ac8bf303a02556
3
+ size 128953
text_encoder/pytorch_model.fp16-00003-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cec283d98160db341000368ba9058d2ab10c68e851fb40fe5de774edf9a70ae
3
+ size 128953
text_encoder/pytorch_model.fp16-00004-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cf56c00ff1b2c95e8d4070a8522ffd9103b8d2d76ff0dbe8824b6070e27fcee
3
+ size 99941
text_encoder/pytorch_model.fp16-00005-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfdffd5e11513aa1b385e266f3560e7c37c0fd1ae0d0d950c884c3da24ed421e
3
+ size 35045
text_encoder/pytorch_model.fp16-00006-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7996ba643da62d4e0f93eaa0866a8af1fad0c5a36ebc0146c91cc3c333288ef5
3
+ size 263097
text_encoder/pytorch_model.fp16-00007-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83bf37a30b8fcfe6aa1beb4664eea741d9efcf75126d085ccf3b80625366df4c
3
+ size 263097
text_encoder/pytorch_model.fp16-00008-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc88ae28392e6927c48569c5bd0c3997019c8c2870ac1433f695fd940cbe005e
3
+ size 100347
text_encoder/pytorch_model.fp16-00009-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:524501cca24031f320980e4ab6dbe977614a3a0519296a98e7040a6c9cf09047
3
+ size 34191
text_encoder/pytorch_model.fp16-00010-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55a5dbba02cc80550fadf36aa7113ccfdc4d2d9e6f8665bdb17664fe9e7298f5
3
+ size 263097
text_encoder/pytorch_model.fp16-00011-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc68946b2e58052265b6fd2ba8609de1b100aa66af0c3e2c67b8c285b04e42a9
3
+ size 263097
text_encoder/pytorch_model.fp16-00012-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcf4545aba827893e0236a49d1617f076c0e23feb7edba0246f09d06eb61e243
3
+ size 100347
text_encoder/pytorch_model.fp16-00013-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c536cb0e84c28ca62932ebc5a819223a99e6534c852084d374a84ef3ea942e5b
3
+ size 34191
text_encoder/pytorch_model.fp16-00014-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd4a7ca8256fa01b63b3d86f72f30195ba78db0627cfcd0153bb28b4156c7a2e
3
+ size 263097
text_encoder/pytorch_model.fp16-00015-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52e146bac50269b00c790b8a7bb6720822e2db15154f9508d8601f59d6c268fc
3
+ size 263097
text_encoder/pytorch_model.fp16-00016-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eea203ebd1fcb6f4088bf19fa51e577ae94067028b76d64dd85a381618555443
3
+ size 100347
text_encoder/pytorch_model.fp16-00017-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9abb7e44b8e5fdad489cf5b2fbb830e80b1b559f002a8f3239a03a104e48012e
3
+ size 34191
text_encoder/pytorch_model.fp16-00018-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3edd89fafc379f394fc3b1eaa842680630c2d5fe04db7a4ed6843e35b24badc5
3
+ size 263097
text_encoder/pytorch_model.fp16-00019-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60128f703795f815fbf40ddb5ba8095ac112ede2e0e248c78801cb4aaa2c2b54
3
+ size 263097
text_encoder/pytorch_model.fp16-00020-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3dff56564702a544b422024d129d9dc17cc3f0c890d2a47425b9accc4938a7c1
3
+ size 100347
text_encoder/pytorch_model.fp16-00021-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfe2bbcdfcf567c0573a55788e0881569363fd14d61e6809656378d37f1d5875
3
+ size 34191
text_encoder/pytorch_model.fp16-00022-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:123a98b31af9affeb5c5ce1b9d578b2c07454041c40ba6008c7bfe374b4588b4
3
+ size 263097
text_encoder/pytorch_model.fp16-00023-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74eca53fb9e9812907443003e419447398338a202f25d74853ae4a85255af4af
3
+ size 263097
text_encoder/pytorch_model.fp16-00024-of-00024.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb2a64a3e9518ad70d0702526aa2677bcf12b00fc8b2528891864fb1c99683e1
3
+ size 1487