Pacific-i64 commited on
Commit
eb7314b
·
verified ·
1 Parent(s): 22b5da4

Upload 6 files

Browse files
config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "hidden_size": 768,
3
+ "num_hidden_layers": 18,
4
+ "intermediate_size": 2416,
5
+ "vocab_size": 32000,
6
+ "num_attention_heads": 12,
7
+ "num_key_value_heads": 4,
8
+ "attention_type": "gqa",
9
+ "attention_dropout": 0.0,
10
+ "use_qk_norm": true,
11
+ "sliding_window": null,
12
+ "max_position_embeddings": 2048,
13
+ "rope_theta": 10000.0,
14
+ "rope_type": "standard",
15
+ "mlp_type": "swiglu",
16
+ "hidden_act": "silu",
17
+ "num_experts": 1,
18
+ "token_frequencies": null,
19
+ "use_mu_guidance": false,
20
+ "use_mu_projection": false,
21
+ "disable_mu_guidance": false,
22
+ "norm_type": "rmsnorm",
23
+ "norm_eps": 1e-06,
24
+ "tie_word_embeddings": true,
25
+ "use_sdpa": true,
26
+ "use_cache": true,
27
+ "initializer_range": 0.02,
28
+ "extra_config": {}
29
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d235b0129657f8dcc30b686adfce5c74e845a49ea8de2e77d87ff246d79abe55
3
+ size 341648672
model_config.yaml ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ attention_dropout: 0.0
2
+ attention_type: gqa
3
+ disable_mu_guidance: false
4
+ extra_config: {}
5
+ hidden_act: silu
6
+ hidden_size: 768
7
+ initializer_range: 0.02
8
+ intermediate_size: 2416
9
+ max_position_embeddings: 2048
10
+ mlp_type: swiglu
11
+ norm_eps: 1.0e-06
12
+ norm_type: rmsnorm
13
+ num_attention_heads: 12
14
+ num_experts: 1
15
+ num_hidden_layers: 18
16
+ num_key_value_heads: 4
17
+ rope_theta: 10000.0
18
+ rope_type: standard
19
+ sliding_window: null
20
+ tie_word_embeddings: true
21
+ token_frequencies: null
22
+ use_cache: true
23
+ use_mu_guidance: false
24
+ use_mu_projection: false
25
+ use_qk_norm: true
26
+ use_sdpa: true
27
+ vocab_size: 32000
special_tokens_map.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "eos_token": "</s>",
4
+ "pad_token": "<pad>",
5
+ "unk_token": "<unk>"
6
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": true,
4
+ "chat_template": "{% if messages[0]['role'] == 'system' %}{{ messages[0]['content'] }}\n\n{% set ns = namespace(messages=messages[1:]) %}{% else %}{% set ns = namespace(messages=messages) %}{% endif %}{% for message in ns.messages %}{% if message['role'] == 'user' %}User: {{ message['content'] }}\n\n{% elif message['role'] == 'assistant' %}Assistant: {{ message['content'] }}{{ eos_token }}\n{% endif %}{% endfor %}{% if add_generation_prompt %}Assistant:{% endif %}",
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "</s>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<pad>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "<s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "3": {
31
+ "content": "<unk>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ }
38
+ },
39
+ "bos_token": "<s>",
40
+ "clean_up_tokenization_spaces": false,
41
+ "eos_token": "</s>",
42
+ "extra_special_tokens": {},
43
+ "model_max_length": 2048,
44
+ "pad_token": "<pad>",
45
+ "tokenizer_class": "PreTrainedTokenizerFast",
46
+ "unk_token": "<unk>"
47
+ }