Aremaki commited on
Commit
1723db8
·
1 Parent(s): 9e68ee6

json files

Browse files
.gitattributes ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model-00001-of-00004.safetensors filter=lfs diff=lfs merge=lfs -text
2
+ model-00002-of-00004.safetensors filter=lfs diff=lfs merge=lfs -text
3
+ model-00003-of-00004.safetensors filter=lfs diff=lfs merge=lfs -text
4
+ model-00004-of-00004.safetensors filter=lfs diff=lfs merge=lfs -text
5
+ optimizer.pt filter=lfs diff=lfs merge=lfs -text
6
+ scheduler.pt filter=lfs diff=lfs merge=lfs -text
7
+ rng_state.pth filter=lfs diff=lfs merge=lfs -text
8
+ candidate_trie.pkl filter=lfs diff=lfs merge=lfs -text
9
+ config.json filter=lfs diff=lfs merge=lfs -text
10
+ generation_config.json filter=lfs diff=lfs merge=lfs -text
11
+ model.safetensors.index.json filter=lfs diff=lfs merge=lfs -text
12
+ special_tokens_map.json filter=lfs diff=lfs merge=lfs -text
13
+ text_to_code.json filter=lfs diff=lfs merge=lfs -text
14
+ tokenizer_config.json filter=lfs diff=lfs merge=lfs -text
15
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
16
+ trainer_state.json filter=lfs diff=lfs merge=lfs -text
candidate_trie.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:096c806cf4b0b3a18a807919deff30ca8fe70faceda8e7cfa5ae3ab7bae4554a
3
+ size 23627786
config.json CHANGED
@@ -1,35 +1,3 @@
1
- {
2
- "architectures": [
3
- "LLamaSynCABEL"
4
- ],
5
- "attention_bias": false,
6
- "attention_dropout": 0.0,
7
- "bos_token_id": 128000,
8
- "dtype": "bfloat16",
9
- "eos_token_id": 128001,
10
- "head_dim": 128,
11
- "hidden_act": "silu",
12
- "hidden_size": 4096,
13
- "initializer_range": 0.02,
14
- "intermediate_size": 14336,
15
- "max_position_embeddings": 8192,
16
- "mlp_bias": false,
17
- "model_type": "llama_syncabel",
18
- "auto_map": {
19
- "AutoConfig": "syncabel.LLamaSynCABELConfig",
20
- "AutoModelForCausalLM": "syncabel.LLamaSynCABEL"
21
- },
22
- "num_attention_heads": 32,
23
- "num_hidden_layers": 32,
24
- "num_key_value_heads": 8,
25
- "pad_token_id": 128001,
26
- "pretraining_tp": 1,
27
- "rms_norm_eps": 1e-05,
28
- "rope_scaling": null,
29
- "rope_theta": 500000.0,
30
- "tie_word_embeddings": false,
31
- "transformers_version": "4.57.1",
32
- "use_cache": true,
33
- "vocab_size": 128257,
34
- "lang": "es"
35
- }
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfe424e6a1783a4a90832e27b440bf2b4fa5e980f73fd2c090852307b0882a25
3
+ size 855
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
generation_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd1490e4e4ab156bedbd15bc636ed214a42424df861d94fe8d548e7c588e1824
3
+ size 157
model.safetensors.index.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c4edf7af872ca8cd250b5ee9871ab0959b6114b66f613c3d1cc0b24acb2c455
3
+ size 23986
special_tokens_map.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4531320e3914ddd27d3858a598c2f9d756055c858b22b6a2f86ef47c61bd6915
3
+ size 587
text_to_code.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c3dacc1f75cc8e61938246e5410a33d61969382039186ea4ef31527c98e0a57
3
+ size 36809012
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4a3d14b25454130fd21476a0277df467e0be8b2ba6785af6f653351b4df60af
3
+ size 17208935
tokenizer_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14e4e1ed2f6dce88035435bb30b8f23e8b644c70dbaae8d6d3c2ddee6a92793c
3
+ size 50826
trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21dea053b71512fee51d0408120fb0b9f9a5b6b9ebe977d0dffd5932d9875315
3
+ size 37541