ToastyPigeon commited on
Commit
d492fac
·
verified ·
1 Parent(s): 1fec294

Training in progress, step 5

Browse files
adapter_config.json CHANGED
@@ -24,13 +24,13 @@
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
- "o_proj",
28
- "down_proj",
29
  "up_proj",
 
 
30
  "k_proj",
31
  "q_proj",
32
- "gate_proj",
33
- "v_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
+ "v_proj",
 
28
  "up_proj",
29
+ "down_proj",
30
+ "o_proj",
31
  "k_proj",
32
  "q_proj",
33
+ "gate_proj"
 
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:95acb38d014584c6b40597db7ee65c8089400fb21322fdadfc4f31ae2a16e3b5
3
- size 456206712
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39d4dd96080c3616aa6d08dae3d76f1b345376cc1812a392813a337b646bfcf0
3
+ size 3140603848
config.json CHANGED
@@ -4,7 +4,7 @@
4
  ],
5
  "attention_dropout": 0.0,
6
  "bos_token_id": 1,
7
- "eos_token_id": 15,
8
  "head_dim": 128,
9
  "hidden_act": "silu",
10
  "hidden_size": 5120,
@@ -37,5 +37,5 @@
37
  "torch_dtype": "bfloat16",
38
  "transformers_version": "4.52.4",
39
  "use_cache": false,
40
- "vocab_size": 131072
41
  }
 
4
  ],
5
  "attention_dropout": 0.0,
6
  "bos_token_id": 1,
7
+ "eos_token_id": 131072,
8
  "head_dim": 128,
9
  "hidden_act": "silu",
10
  "hidden_size": 5120,
 
37
  "torch_dtype": "bfloat16",
38
  "transformers_version": "4.52.4",
39
  "use_cache": false,
40
+ "vocab_size": 131074
41
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4b4c8fcd33487a449c07f423d47adb035bba8347ccf13eb074b4d1fef8acf919
3
- size 17078288
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2fa2956478eaa353c6c4b1f47fdd6868cce6075e52e169c35ae8bd28524e7a8
3
+ size 17078668
tokenizer_config.json CHANGED
@@ -116,7 +116,7 @@
116
  "special": true
117
  },
118
  "14": {
119
- "content": "<|im_start|>",
120
  "lstrip": false,
121
  "normalized": false,
122
  "rstrip": false,
@@ -124,7 +124,7 @@
124
  "special": true
125
  },
126
  "15": {
127
- "content": "<|im_end|>",
128
  "lstrip": false,
129
  "normalized": false,
130
  "rstrip": false,
@@ -8002,6 +8002,22 @@
8002
  "rstrip": false,
8003
  "single_word": false,
8004
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8005
  }
8006
  },
8007
  "bos_token": "<s>",
 
116
  "special": true
117
  },
118
  "14": {
119
+ "content": "<SPECIAL_14>",
120
  "lstrip": false,
121
  "normalized": false,
122
  "rstrip": false,
 
124
  "special": true
125
  },
126
  "15": {
127
+ "content": "<SPECIAL_15>",
128
  "lstrip": false,
129
  "normalized": false,
130
  "rstrip": false,
 
8002
  "rstrip": false,
8003
  "single_word": false,
8004
  "special": true
8005
+ },
8006
+ "131072": {
8007
+ "content": "<|im_end|>",
8008
+ "lstrip": false,
8009
+ "normalized": false,
8010
+ "rstrip": false,
8011
+ "single_word": false,
8012
+ "special": true
8013
+ },
8014
+ "131073": {
8015
+ "content": "<|im_start|>",
8016
+ "lstrip": false,
8017
+ "normalized": false,
8018
+ "rstrip": false,
8019
+ "single_word": false,
8020
+ "special": true
8021
  }
8022
  },
8023
  "bos_token": "<s>",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:00680aba3e8386e5d6534a13bd4e116c275cde571d5790cdf1205c8fcd88cefb
3
  size 9489
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b863d975b84ce444ea5c542d9fabff254c1ce04e88ce402c0259419976063185
3
  size 9489