voidful commited on
Commit
c656cd3
·
verified ·
1 Parent(s): b8af831

Training in progress, epoch 3

Browse files
adapter_config.json CHANGED
@@ -1,9 +1,12 @@
1
  {
 
2
  "alpha_pattern": {},
 
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "voidful/llm-codec",
5
  "bias": "none",
6
  "corda_config": null,
 
7
  "eva_config": null,
8
  "exclude_modules": null,
9
  "fan_in_fan_out": false,
@@ -20,20 +23,24 @@
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
 
 
23
  "r": 64,
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
- "gate_proj",
28
- "k_proj",
29
  "q_proj",
30
- "up_proj",
31
  "o_proj",
 
 
32
  "v_proj",
33
- "down_proj"
 
34
  ],
 
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
37
  "use_dora": false,
 
38
  "use_rslora": false
39
  }
 
1
  {
2
+ "alora_invocation_tokens": null,
3
  "alpha_pattern": {},
4
+ "arrow_config": null,
5
  "auto_mapping": null,
6
  "base_model_name_or_path": "voidful/llm-codec",
7
  "bias": "none",
8
  "corda_config": null,
9
+ "ensure_weight_tying": false,
10
  "eva_config": null,
11
  "exclude_modules": null,
12
  "fan_in_fan_out": false,
 
23
  "megatron_core": "megatron.core",
24
  "modules_to_save": null,
25
  "peft_type": "LORA",
26
+ "peft_version": "0.18.0",
27
+ "qalora_group_size": 16,
28
  "r": 64,
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
 
 
32
  "q_proj",
 
33
  "o_proj",
34
+ "k_proj",
35
+ "down_proj",
36
  "v_proj",
37
+ "up_proj",
38
+ "gate_proj"
39
  ],
40
+ "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
42
  "trainable_token_indices": null,
43
  "use_dora": false,
44
+ "use_qalora": false,
45
  "use_rslora": false
46
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ec670faef160266af110492dd83512f76fd0c9d4d739ebbee32a75847acc420
3
  size 528550256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a16703691beb3920653d04fcd565151a9415fcd15b74c1c014b7c25cbad0bbf
3
  size 528550256
tokenizer_config.json CHANGED
@@ -164071,13 +164071,9 @@
164071
  "eos_token": "<|im_end|>",
164072
  "errors": "replace",
164073
  "extra_special_tokens": {},
164074
- "max_length": 4096,
164075
  "model_max_length": 1010000,
164076
  "pad_token": "<|endoftext|>",
164077
  "split_special_tokens": false,
164078
- "stride": 0,
164079
  "tokenizer_class": "Qwen2Tokenizer",
164080
- "truncation_side": "right",
164081
- "truncation_strategy": "longest_first",
164082
  "unk_token": null
164083
  }
 
164071
  "eos_token": "<|im_end|>",
164072
  "errors": "replace",
164073
  "extra_special_tokens": {},
 
164074
  "model_max_length": 1010000,
164075
  "pad_token": "<|endoftext|>",
164076
  "split_special_tokens": false,
 
164077
  "tokenizer_class": "Qwen2Tokenizer",
 
 
164078
  "unk_token": null
164079
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:09ef793003aa3f31e0fa9d57c9998371626de0241ae516edb6c9795266860010
3
  size 5841
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:086a4fbc7538a577fb092ba511f7e1a4d6b3867e138da19ed05f63656337f331
3
  size 5841