Victor1306 commited on
Commit
ab46fe4
·
verified ·
1 Parent(s): c92c556

Upload folder using huggingface_hub

Browse files
config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "cls",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "hidden_activation": "gelu",
20
+ "hidden_size": 448,
21
+ "initializer_cutoff_factor": 2.0,
22
+ "initializer_range": 0.02,
23
+ "intermediate_size": 896,
24
+ "local_attention": 128,
25
+ "local_rope_theta": 10000.0,
26
+ "max_position_embeddings": 8192,
27
+ "mlp_bias": false,
28
+ "mlp_dropout": 0.0,
29
+ "model_type": "modernbert",
30
+ "norm_bias": false,
31
+ "norm_eps": 1e-05,
32
+ "num_attention_heads": 14,
33
+ "num_hidden_layers": 24,
34
+ "pad_token_id": 0,
35
+ "repad_logits_with_grad": false,
36
+ "sep_token_id": 50282,
37
+ "sparse_pred_ignore_index": -100,
38
+ "sparse_prediction": false,
39
+ "torch_dtype": "float32",
40
+ "transformers_version": "4.52.4",
41
+ "vocab_size": 29
42
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:636de3ccf8b72ba4c6338a41f45341fc06d52e1e22a165e9e0245bb1a2123746
3
+ size 192829632
special_tokens_map.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "eos_token": "</s>",
3
+ "mask_token": "<mask>",
4
+ "pad_token": "<pad>",
5
+ "unk_token": "<unk>"
6
+ }
tokenizer.json ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
5
+ "added_tokens": [],
6
+ "normalizer": null,
7
+ "pre_tokenizer": {
8
+ "type": "Whitespace"
9
+ },
10
+ "post_processor": null,
11
+ "decoder": null,
12
+ "model": {
13
+ "type": "WordLevel",
14
+ "vocab": {
15
+ "<pad>": 0,
16
+ "</s>": 1,
17
+ "<unk>": 2,
18
+ "A": 3,
19
+ "L": 4,
20
+ "G": 5,
21
+ "V": 6,
22
+ "S": 7,
23
+ "R": 8,
24
+ "E": 9,
25
+ "D": 10,
26
+ "T": 11,
27
+ "I": 12,
28
+ "P": 13,
29
+ "K": 14,
30
+ "F": 15,
31
+ "Q": 16,
32
+ "N": 17,
33
+ "Y": 18,
34
+ "M": 19,
35
+ "H": 20,
36
+ "W": 21,
37
+ "C": 22,
38
+ "X": 23,
39
+ "B": 24,
40
+ "O": 25,
41
+ "U": 26,
42
+ "Z": 27,
43
+ "<mask>": 28
44
+ },
45
+ "unk_token": "<unk>"
46
+ }
47
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "model_max_length": 1000000000000000000
3
+ }
vocab.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</s>": 1,
3
+ "<mask>": 28,
4
+ "<pad>": 0,
5
+ "<unk>": 2,
6
+ "A": 3,
7
+ "B": 24,
8
+ "C": 22,
9
+ "D": 10,
10
+ "E": 9,
11
+ "F": 15,
12
+ "G": 5,
13
+ "H": 20,
14
+ "I": 12,
15
+ "K": 14,
16
+ "L": 4,
17
+ "M": 19,
18
+ "N": 17,
19
+ "O": 25,
20
+ "P": 13,
21
+ "Q": 16,
22
+ "R": 8,
23
+ "S": 7,
24
+ "T": 11,
25
+ "U": 26,
26
+ "V": 6,
27
+ "W": 21,
28
+ "X": 23,
29
+ "Y": 18,
30
+ "Z": 27
31
+ }