error577 commited on
Commit
b7b85f0
·
verified ·
1 Parent(s): 4c41552

End of training

Browse files
README.md CHANGED
@@ -66,7 +66,7 @@ lora_model_dir: null
66
  lora_r: 32
67
  lora_target_linear: true
68
  lr_scheduler: cosine
69
- max_steps: 20
70
  micro_batch_size: 1
71
  mlflow_experiment_name: /tmp/bd684c24fae39600_train_data.json
72
  model_type: AutoModelForCausalLM
@@ -103,7 +103,7 @@ xformers_attention: null
103
 
104
  This model is a fine-tuned version of [unsloth/Qwen2.5-Coder-7B-Instruct](https://huggingface.co/unsloth/Qwen2.5-Coder-7B-Instruct) on the None dataset.
105
  It achieves the following results on the evaluation set:
106
- - Loss: 1.0922
107
 
108
  ## Model description
109
 
@@ -131,17 +131,16 @@ The following hyperparameters were used during training:
131
  - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
132
  - lr_scheduler_type: cosine
133
  - lr_scheduler_warmup_steps: 10
134
- - training_steps: 20
135
 
136
  ### Training results
137
 
138
  | Training Loss | Epoch | Step | Validation Loss |
139
  |:-------------:|:------:|:----:|:---------------:|
140
  | 1.4698 | 0.0015 | 1 | 1.7292 |
141
- | 1.2989 | 0.0075 | 5 | 1.7222 |
142
- | 1.1088 | 0.0150 | 10 | 1.5535 |
143
- | 0.8043 | 0.0225 | 15 | 1.2118 |
144
- | 0.7881 | 0.0300 | 20 | 1.0922 |
145
 
146
 
147
  ### Framework versions
 
66
  lora_r: 32
67
  lora_target_linear: true
68
  lr_scheduler: cosine
69
+ max_steps: 50
70
  micro_batch_size: 1
71
  mlflow_experiment_name: /tmp/bd684c24fae39600_train_data.json
72
  model_type: AutoModelForCausalLM
 
103
 
104
  This model is a fine-tuned version of [unsloth/Qwen2.5-Coder-7B-Instruct](https://huggingface.co/unsloth/Qwen2.5-Coder-7B-Instruct) on the None dataset.
105
  It achieves the following results on the evaluation set:
106
+ - Loss: 0.0745
107
 
108
  ## Model description
109
 
 
131
  - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
132
  - lr_scheduler_type: cosine
133
  - lr_scheduler_warmup_steps: 10
134
+ - training_steps: 50
135
 
136
  ### Training results
137
 
138
  | Training Loss | Epoch | Step | Validation Loss |
139
  |:-------------:|:------:|:----:|:---------------:|
140
  | 1.4698 | 0.0015 | 1 | 1.7292 |
141
+ | 1.2364 | 0.0195 | 13 | 1.3321 |
142
+ | 0.289 | 0.0390 | 26 | 0.3452 |
143
+ | 0.0635 | 0.0584 | 39 | 0.0745 |
 
144
 
145
 
146
  ### Framework versions
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "k_proj",
24
  "down_proj",
25
- "o_proj",
26
- "v_proj",
27
- "q_proj",
28
  "up_proj",
29
- "gate_proj"
 
 
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "down_proj",
 
 
 
24
  "up_proj",
25
+ "k_proj",
26
+ "v_proj",
27
+ "o_proj",
28
+ "gate_proj",
29
+ "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68604427cd083b580fc3c5d4d8958b7a287c470ad87b2a1ac645c43cc13c6bf2
3
  size 323103018
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5f8ff9917efa786ab1cf4cf474ebfc8a53543c5bade01f9c1d46b811fee9554
3
  size 323103018
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19cbf4ecbfce042252e641022bd504ddfce76d4ae8beaef0e80bba513676731b
3
  size 323014168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78102a43db16e692d1ea6a566807b28ac716dcf4f4908683065afe8e41fc9a5f
3
  size 323014168
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2582b8260dcd9ba74e4729f3dca3eddc9202355c03409cb87a0c310e92a237dd
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:004d3e011643fbfcbb780589c7d95d33b23c819b8403b6d5a07ab9641302271b
3
  size 6776