Ba2han commited on
Commit
490cfed
·
verified ·
1 Parent(s): 13d14ae

Training in progress, step 5390

Browse files
README.md CHANGED
@@ -4,9 +4,9 @@ library_name: transformers
4
  model_name: m-lora-6
5
  tags:
6
  - generated_from_trainer
7
- - trl
8
  - sft
9
  - unsloth
 
10
  licence: license
11
  ---
12
 
@@ -28,7 +28,7 @@ print(output["generated_text"])
28
 
29
  ## Training procedure
30
 
31
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/batuhan409/huggingface/runs/4bjr5n1w)
32
 
33
 
34
  This model was trained with SFT.
 
4
  model_name: m-lora-6
5
  tags:
6
  - generated_from_trainer
 
7
  - sft
8
  - unsloth
9
+ - trl
10
  licence: license
11
  ---
12
 
 
28
 
29
  ## Training procedure
30
 
31
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/batuhan409/huggingface/runs/ofmmlzwd)
32
 
33
 
34
  This model was trained with SFT.
adapter_config.json CHANGED
@@ -19,7 +19,7 @@
19
  "loftq_config": {},
20
  "lora_alpha": 32,
21
  "lora_bias": false,
22
- "lora_dropout": 0,
23
  "megatron_config": null,
24
  "megatron_core": "megatron.core",
25
  "modules_to_save": null,
@@ -29,13 +29,13 @@
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
 
32
  "k_proj",
 
33
  "down_proj",
34
  "gate_proj",
35
- "v_proj",
36
- "o_proj",
37
  "up_proj",
38
- "q_proj"
39
  ],
40
  "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
 
19
  "loftq_config": {},
20
  "lora_alpha": 32,
21
  "lora_bias": false,
22
+ "lora_dropout": 0.2,
23
  "megatron_config": null,
24
  "megatron_core": "megatron.core",
25
  "modules_to_save": null,
 
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
+ "o_proj",
33
  "k_proj",
34
+ "q_proj",
35
  "down_proj",
36
  "gate_proj",
 
 
37
  "up_proj",
38
+ "v_proj"
39
  ],
40
  "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cdbd7b502100ce688ea1e8b4881ae33cae07b801288f8dff41c18019d14ca68a
3
  size 218134048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8c38f3f03e4b138df55fe847a20e21813f878ab777c9c72eb5ecd3b74b0a4a2
3
  size 218134048
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9322b7231b1056db4a62b1e1e0ae34c46cfa3ee0144e4d6ec9a35bf905c96d8c
3
  size 6225
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7b9120eb252b286a888ff69fcd7cf62d677791ced297964bb7381a31035c3c5
3
  size 6225