Ba2han commited on
Commit
26007b7
·
verified ·
1 Parent(s): 26ebdd6

Training in progress, step 362

Browse files
README.md CHANGED
@@ -5,8 +5,8 @@ model_name: test-gemini
5
  tags:
6
  - generated_from_trainer
7
  - trl
8
- - sft
9
  - unsloth
 
10
  licence: license
11
  ---
12
 
@@ -28,7 +28,7 @@ print(output["generated_text"])
28
 
29
  ## Training procedure
30
 
31
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/batuhan409/huggingface/runs/43fs3x2p)
32
 
33
 
34
  This model was trained with SFT.
 
5
  tags:
6
  - generated_from_trainer
7
  - trl
 
8
  - unsloth
9
+ - sft
10
  licence: license
11
  ---
12
 
 
28
 
29
  ## Training procedure
30
 
31
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/batuhan409/huggingface/runs/k89bvfrt)
32
 
33
 
34
  This model was trained with SFT.
adapter_config.json CHANGED
@@ -20,7 +20,7 @@
20
  "layers_pattern": null,
21
  "layers_to_transform": null,
22
  "loftq_config": {},
23
- "lora_alpha": 256,
24
  "lora_bias": false,
25
  "lora_dropout": 0,
26
  "megatron_config": null,
@@ -35,11 +35,11 @@
35
  "target_modules": [
36
  "k_proj",
37
  "down_proj",
38
- "q_proj",
 
39
  "o_proj",
40
  "up_proj",
41
- "gate_proj",
42
- "v_proj"
43
  ],
44
  "target_parameters": null,
45
  "task_type": "CAUSAL_LM",
 
20
  "layers_pattern": null,
21
  "layers_to_transform": null,
22
  "loftq_config": {},
23
+ "lora_alpha": 128,
24
  "lora_bias": false,
25
  "lora_dropout": 0,
26
  "megatron_config": null,
 
35
  "target_modules": [
36
  "k_proj",
37
  "down_proj",
38
+ "gate_proj",
39
+ "v_proj",
40
  "o_proj",
41
  "up_proj",
42
+ "q_proj"
 
43
  ],
44
  "target_parameters": null,
45
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b3ce31ba0e668c10d07dfa8a7e2fd348b76585474273a196e29570d05549cf56
3
  size 1347775120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52aa2324035c1292bf2fa39f648f91ac5671d3d542c9e4e2b5783d5972371b74
3
  size 1347775120
tokenizer_config.json CHANGED
@@ -11,7 +11,7 @@
11
  ],
12
  "model_max_length": 131072,
13
  "model_specific_special_tokens": {},
14
- "pad_token": "<|reserved_special_token_250|>",
15
  "padding_side": "right",
16
  "tokenizer_class": "TokenizersBackend"
17
  }
 
11
  ],
12
  "model_max_length": 131072,
13
  "model_specific_special_tokens": {},
14
+ "pad_token": "<|eot_id|>",
15
  "padding_side": "right",
16
  "tokenizer_class": "TokenizersBackend"
17
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c30167f40d7b7a967ef44dde73679b16467d012fe5bd5a7b809f4da605df0cb
3
  size 5713
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35428c9ecf443b790ca66a1a03447a6eeb723685fc36c79a6b97241be28d4f9c
3
  size 5713