error577 commited on
Commit
468e483
·
verified ·
1 Parent(s): 223c4f7

Training in progress, step 100

Browse files
README.md CHANGED
@@ -64,7 +64,7 @@ lora_fan_in_fan_out: null
64
  lora_model_dir: null
65
  lora_r: 32
66
  lora_target_linear: true
67
- loraplus_lr_ratio: 16
68
  lr_scheduler: cosine
69
  micro_batch_size: 1
70
  mlflow_experiment_name: /tmp/8e23ab7f3136aade_train_data.json
@@ -103,7 +103,7 @@ xformers_attention: null
103
 
104
  This model is a fine-tuned version of [unsloth/Qwen2.5-3B](https://huggingface.co/unsloth/Qwen2.5-3B) on the None dataset.
105
  It achieves the following results on the evaluation set:
106
- - Loss: 0.8736
107
 
108
  ## Model description
109
 
@@ -139,12 +139,15 @@ No additional optimizer arguments
139
  | Training Loss | Epoch | Step | Validation Loss |
140
  |:-------------:|:------:|:----:|:---------------:|
141
  | 0.5376 | 0.0002 | 1 | 1.1038 |
142
- | 1.0685 | 0.0164 | 100 | 0.8534 |
143
- | 0.9163 | 0.0328 | 200 | 0.8613 |
144
- | 0.9678 | 0.0492 | 300 | 0.8594 |
145
- | 0.9531 | 0.0656 | 400 | 0.8629 |
146
- | 1.0449 | 0.0820 | 500 | 0.8701 |
147
- | 0.8456 | 0.0985 | 600 | 0.8736 |
 
 
 
148
 
149
 
150
  ### Framework versions
 
64
  lora_model_dir: null
65
  lora_r: 32
66
  lora_target_linear: true
67
+ loraplus_lr_ratio: 8
68
  lr_scheduler: cosine
69
  micro_batch_size: 1
70
  mlflow_experiment_name: /tmp/8e23ab7f3136aade_train_data.json
 
103
 
104
  This model is a fine-tuned version of [unsloth/Qwen2.5-3B](https://huggingface.co/unsloth/Qwen2.5-3B) on the None dataset.
105
  It achieves the following results on the evaluation set:
106
+ - Loss: 0.8224
107
 
108
  ## Model description
109
 
 
139
  | Training Loss | Epoch | Step | Validation Loss |
140
  |:-------------:|:------:|:----:|:---------------:|
141
  | 0.5376 | 0.0002 | 1 | 1.1038 |
142
+ | 1.0376 | 0.0164 | 100 | 0.8196 |
143
+ | 0.8848 | 0.0328 | 200 | 0.8143 |
144
+ | 0.9346 | 0.0492 | 300 | 0.8134 |
145
+ | 0.8966 | 0.0656 | 400 | 0.8115 |
146
+ | 1.008 | 0.0820 | 500 | 0.8160 |
147
+ | 0.8225 | 0.0985 | 600 | 0.8154 |
148
+ | 0.857 | 0.1149 | 700 | 0.8197 |
149
+ | 1.0713 | 0.1313 | 800 | 0.8197 |
150
+ | 1.0041 | 0.1477 | 900 | 0.8224 |
151
 
152
 
153
  ### Framework versions
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "k_proj",
24
  "o_proj",
25
- "gate_proj",
26
- "v_proj",
27
  "q_proj",
 
28
  "down_proj",
29
- "up_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "up_proj",
24
  "o_proj",
 
 
25
  "q_proj",
26
+ "gate_proj",
27
  "down_proj",
28
+ "v_proj",
29
+ "k_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c21555a7c41429f66d7bb15398b4bd28d52aef56734104fbcf8126caec3ffeb2
3
  size 239650666
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2b97e7367f463598722838ea3e789eafccc2a2bf2de7bb47ba772e4f6b63839
3
  size 239650666
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:df2c88e441560b7d30a0c68235fe450bd704794674e100b7ef8872d879e42c39
3
  size 239536272
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97e0aa8302f740897c3fc855e7ccb97c0f8d2873de5a0776adfdd8f5e0139e68
3
  size 239536272
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6f892413ee6a33391e91a66fafb17b7974d3c90c2418e9ad8556ec130bd8439
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:969d207d1f6c0d67768fcd41b7675980b9dc88c38b5029cd598ebef348814045
3
  size 6776