apepkuss79 commited on
Commit
22186c6
·
verified ·
1 Parent(s): caf82a2

Update models

Browse files
.gitattributes CHANGED
@@ -33,3 +33,16 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ embeddinggemma-300m-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ embeddinggemma-300m-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ embeddinggemma-300m-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ embeddinggemma-300m-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ embeddinggemma-300m-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ embeddinggemma-300m-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ embeddinggemma-300m-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ embeddinggemma-300m-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ embeddinggemma-300m-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ embeddinggemma-300m-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ embeddinggemma-300m-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ embeddinggemma-300m-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
48
+ embeddinggemma-300m-f16.gguf filter=lfs diff=lfs merge=lfs -text
config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_sliding_window_pattern": 6,
3
+ "architectures": [
4
+ "Gemma3TextModel"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "attn_logit_softcapping": null,
9
+ "bos_token_id": 2,
10
+ "dtype": "float32",
11
+ "eos_token_id": 1,
12
+ "final_logit_softcapping": null,
13
+ "head_dim": 256,
14
+ "hidden_activation": "gelu_pytorch_tanh",
15
+ "hidden_size": 768,
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 1152,
18
+ "layer_types": [
19
+ "sliding_attention",
20
+ "sliding_attention",
21
+ "sliding_attention",
22
+ "sliding_attention",
23
+ "sliding_attention",
24
+ "full_attention",
25
+ "sliding_attention",
26
+ "sliding_attention",
27
+ "sliding_attention",
28
+ "sliding_attention",
29
+ "sliding_attention",
30
+ "full_attention",
31
+ "sliding_attention",
32
+ "sliding_attention",
33
+ "sliding_attention",
34
+ "sliding_attention",
35
+ "sliding_attention",
36
+ "full_attention",
37
+ "sliding_attention",
38
+ "sliding_attention",
39
+ "sliding_attention",
40
+ "sliding_attention",
41
+ "sliding_attention",
42
+ "full_attention"
43
+ ],
44
+ "max_position_embeddings": 2048,
45
+ "model_type": "gemma3_text",
46
+ "num_attention_heads": 3,
47
+ "num_hidden_layers": 24,
48
+ "num_key_value_heads": 1,
49
+ "pad_token_id": 0,
50
+ "query_pre_attn_scalar": 256,
51
+ "rms_norm_eps": 1e-06,
52
+ "rope_local_base_freq": 10000.0,
53
+ "rope_scaling": null,
54
+ "rope_theta": 1000000.0,
55
+ "sliding_window": 512,
56
+ "transformers_version": "4.57.0.dev0",
57
+ "use_bidirectional_attention": true,
58
+ "use_cache": true,
59
+ "vocab_size": 262144
60
+ }
embeddinggemma-300m-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ea1444eb32a100903684e2f587e0fd71151f88ffc0087264625b23bd8b14657
3
+ size 212209632
embeddinggemma-300m-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef60f84fa14d25fe9e0c40b5adea5413712eff34f2d4eafa2ddb0ca5fecc6cec
3
+ size 227287008
embeddinggemma-300m-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13724b207f62d393ef7bd00bacc80204a062f6f680fdf2e5c07dbd068d872f8a
3
+ size 223705056
embeddinggemma-300m-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:876e9a13b3e55caa41c64e3c32ccf80c54f98302fed76c77ac39665b1e0f04a0
3
+ size 218439648
embeddinggemma-300m-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b6fb167834b37a5296122f373c04ab1ca91a5d79ba8f890516e654acd49e0fb
3
+ size 229093344
embeddinggemma-300m-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2decb24229f3876b66edee3f0c825edcf3ac46206819ec51aff29fd97e8ce529
3
+ size 236337120
embeddinggemma-300m-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f419f42341f33fea445cee95c0966c2dc1b849e35d656ea8496fcfef085e05de
3
+ size 232011744
embeddinggemma-300m-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a897dd6ca7fc32c008b81ef7259bcc058c2ec1ca1dc20effaf0f4dda260016c
3
+ size 241774560
embeddinggemma-300m-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13bed4b96c6ea13717d8ff57f177e43300c7674f8be93bb34f4b7f2581ff5c16
3
+ size 246732768
embeddinggemma-300m-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e49639f7416da0e01ffb04df2ac63aec6eaa400a3b61c63ad51a7486c321cf5
3
+ size 243101664
embeddinggemma-300m-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d6a91f5ae2c5043680c6cbfd506390bebeed8a84c18c2d31b9d9a5cd90f741e
3
+ size 260390880
embeddinggemma-300m-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2a9e50d318b59f03d3f0e256cc9eb4092f4793c155a7ffbebb45fa034b5fded
3
+ size 328576992
embeddinggemma-300m-f16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6d9565ffd73075f0577628a97bbfa0f6345076b4273626dc03f06dcf3318122
3
+ size 612429792