error577 commited on
Commit
8e9d9ba
·
verified ·
1 Parent(s): df5a980

End of training

Browse files
README.md CHANGED
@@ -66,11 +66,11 @@ lora_model_dir: null
66
  lora_r: 32
67
  lora_target_linear: true
68
  lr_scheduler: cosine
69
- max_steps: 100
70
  micro_batch_size: 1
71
  mlflow_experiment_name: /tmp/0fe9be330b40f25d_train_data.json
72
  model_type: AutoModelForCausalLM
73
- num_epochs: 1
74
  optimizer: adamw_bnb_8bit
75
  output_dir: miner_id_24
76
  pad_to_sequence_len: true
@@ -103,7 +103,7 @@ xformers_attention: null
103
 
104
  This model is a fine-tuned version of [Qwen/Qwen2.5-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct) on the None dataset.
105
  It achieves the following results on the evaluation set:
106
- - Loss: 1.9689
107
 
108
  ## Model description
109
 
@@ -131,13 +131,17 @@ The following hyperparameters were used during training:
131
  - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
132
  - lr_scheduler_type: cosine
133
  - lr_scheduler_warmup_steps: 10
134
- - training_steps: 100
135
 
136
  ### Training results
137
 
138
  | Training Loss | Epoch | Step | Validation Loss |
139
  |:-------------:|:------:|:----:|:---------------:|
140
- | 1.6039 | 0.0113 | 100 | 1.9689 |
 
 
 
 
141
 
142
 
143
  ### Framework versions
 
66
  lora_r: 32
67
  lora_target_linear: true
68
  lr_scheduler: cosine
69
+ max_steps: 500
70
  micro_batch_size: 1
71
  mlflow_experiment_name: /tmp/0fe9be330b40f25d_train_data.json
72
  model_type: AutoModelForCausalLM
73
+ num_epochs: 4
74
  optimizer: adamw_bnb_8bit
75
  output_dir: miner_id_24
76
  pad_to_sequence_len: true
 
103
 
104
  This model is a fine-tuned version of [Qwen/Qwen2.5-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct) on the None dataset.
105
  It achieves the following results on the evaluation set:
106
+ - Loss: 1.0580
107
 
108
  ## Model description
109
 
 
131
  - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
132
  - lr_scheduler_type: cosine
133
  - lr_scheduler_warmup_steps: 10
134
+ - training_steps: 500
135
 
136
  ### Training results
137
 
138
  | Training Loss | Epoch | Step | Validation Loss |
139
  |:-------------:|:------:|:----:|:---------------:|
140
+ | 5.4801 | 0.0001 | 1 | 6.2892 |
141
+ | 1.4237 | 0.0142 | 125 | 1.3264 |
142
+ | 0.953 | 0.0284 | 250 | 1.1498 |
143
+ | 0.8279 | 0.0425 | 375 | 1.0727 |
144
+ | 1.0675 | 0.0567 | 500 | 1.0580 |
145
 
146
 
147
  ### Framework versions
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "gate_proj",
24
- "q_proj",
25
- "v_proj",
26
  "o_proj",
27
  "k_proj",
 
28
  "up_proj",
29
- "down_proj"
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
23
  "o_proj",
24
  "k_proj",
25
+ "down_proj",
26
  "up_proj",
27
+ "gate_proj",
28
+ "q_proj",
29
+ "v_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1d7fe1df43ac934e7263717d724f4ce2d3389c91b06e7f11e0b38e8f1f4d4209
3
  size 70506570
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3784fa22e586a753a5a30e8dd78c3a7f19b80e00299e7e20aa257775efd588c
3
  size 70506570
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:82281a0105057b4df432900a373bdbb0234a05bf0b3bc9f1f5106eb71273794a
3
  size 70430032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e78bbe4c88e10e5a49e4b7812347a0bc571d773dac6b5888ea1bd69c52f84b16
3
  size 70430032
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7e3b00705b00812815e695fb883fc4148f6c255aedbce9f8d586fee3c83d859e
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03faafebfe7e7e6a881ee31a6d1c8c276772d4b35b0877e4552974cf0e391a70
3
  size 6776