FatCat87 commited on
Commit
b69e640
·
verified ·
1 Parent(s): 5c802b9

End of training

Browse files
Files changed (2) hide show
  1. README.md +19 -20
  2. adapter_model.bin +2 -2
README.md CHANGED
@@ -1,12 +1,12 @@
1
  ---
2
- license: mit
3
  library_name: peft
4
  tags:
5
  - axolotl
6
  - generated_from_trainer
7
- base_model: princeton-nlp/gemma-2-9b-it-SimPO
8
  model-index:
9
- - name: ea995dc6-2b84-41b3-ac45-9d80fc8d62a8
10
  results: []
11
  ---
12
 
@@ -19,19 +19,19 @@ should probably proofread and complete it, then remove this comment. -->
19
  axolotl version: `0.4.1`
20
  ```yaml
21
  adapter: lora
22
- base_model: princeton-nlp/gemma-2-9b-it-SimPO
23
  bf16: auto
24
  datasets:
25
  - data_files:
26
- - 0a8e1ed234d341f6_train_data.json
27
  ds_type: json
28
  format: custom
29
- path: 0a8e1ed234d341f6_train_data.json
30
  type:
31
  field: null
32
- field_input: num
33
- field_instruction: title_main
34
- field_output: texte
35
  field_system: null
36
  format: null
37
  no_input_format: null
@@ -51,7 +51,7 @@ fsdp_config: null
51
  gradient_accumulation_steps: 4
52
  gradient_checkpointing: true
53
  group_by_length: false
54
- hub_model_id: FatCat87/ea995dc6-2b84-41b3-ac45-9d80fc8d62a8
55
  learning_rate: 0.0002
56
  load_in_4bit: false
57
  load_in_8bit: true
@@ -82,9 +82,9 @@ val_set_size: 0.1
82
  wandb_entity: fatcat87-taopanda
83
  wandb_log_model: null
84
  wandb_mode: online
85
- wandb_name: ea995dc6-2b84-41b3-ac45-9d80fc8d62a8
86
  wandb_project: subnet56
87
- wandb_runid: ea995dc6-2b84-41b3-ac45-9d80fc8d62a8
88
  wandb_watch: null
89
  warmup_ratio: 0.05
90
  weight_decay: 0.0
@@ -94,12 +94,12 @@ xformers_attention: null
94
 
95
  </details><br>
96
 
97
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/fatcat87-taopanda/subnet56/runs/1908w8y2)
98
- # ea995dc6-2b84-41b3-ac45-9d80fc8d62a8
99
 
100
- This model is a fine-tuned version of [princeton-nlp/gemma-2-9b-it-SimPO](https://huggingface.co/princeton-nlp/gemma-2-9b-it-SimPO) on the None dataset.
101
  It achieves the following results on the evaluation set:
102
- - Loss: 1.5247
103
 
104
  ## Model description
105
 
@@ -135,10 +135,9 @@ The following hyperparameters were used during training:
135
 
136
  | Training Loss | Epoch | Step | Validation Loss |
137
  |:-------------:|:------:|:----:|:---------------:|
138
- | 2.0061 | 0.0571 | 1 | 1.9442 |
139
- | 1.5778 | 0.2857 | 5 | 1.5892 |
140
- | 1.5005 | 0.5714 | 10 | 1.5397 |
141
- | 1.4559 | 0.8571 | 15 | 1.5247 |
142
 
143
 
144
  ### Framework versions
 
1
  ---
2
+ license: apache-2.0
3
  library_name: peft
4
  tags:
5
  - axolotl
6
  - generated_from_trainer
7
+ base_model: Qwen/Qwen2.5-Math-7B-Instruct
8
  model-index:
9
+ - name: d51aa98d-e0f3-4aaf-ae00-e8da30f740ee
10
  results: []
11
  ---
12
 
 
19
  axolotl version: `0.4.1`
20
  ```yaml
21
  adapter: lora
22
+ base_model: Qwen/Qwen2.5-Math-7B-Instruct
23
  bf16: auto
24
  datasets:
25
  - data_files:
26
+ - a9dedf98b14b8e66_train_data.json
27
  ds_type: json
28
  format: custom
29
+ path: a9dedf98b14b8e66_train_data.json
30
  type:
31
  field: null
32
+ field_input: errors
33
+ field_instruction: original_text
34
+ field_output: correct_text
35
  field_system: null
36
  format: null
37
  no_input_format: null
 
51
  gradient_accumulation_steps: 4
52
  gradient_checkpointing: true
53
  group_by_length: false
54
+ hub_model_id: FatCat87/d51aa98d-e0f3-4aaf-ae00-e8da30f740ee
55
  learning_rate: 0.0002
56
  load_in_4bit: false
57
  load_in_8bit: true
 
82
  wandb_entity: fatcat87-taopanda
83
  wandb_log_model: null
84
  wandb_mode: online
85
+ wandb_name: d51aa98d-e0f3-4aaf-ae00-e8da30f740ee
86
  wandb_project: subnet56
87
+ wandb_runid: d51aa98d-e0f3-4aaf-ae00-e8da30f740ee
88
  wandb_watch: null
89
  warmup_ratio: 0.05
90
  weight_decay: 0.0
 
94
 
95
  </details><br>
96
 
97
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/fatcat87-taopanda/subnet56/runs/cmgp1i7e)
98
+ # d51aa98d-e0f3-4aaf-ae00-e8da30f740ee
99
 
100
+ This model is a fine-tuned version of [Qwen/Qwen2.5-Math-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Math-7B-Instruct) on the None dataset.
101
  It achieves the following results on the evaluation set:
102
+ - Loss: 1.3265
103
 
104
  ## Model description
105
 
 
135
 
136
  | Training Loss | Epoch | Step | Validation Loss |
137
  |:-------------:|:------:|:----:|:---------------:|
138
+ | 1.7936 | 0.1818 | 1 | 1.9329 |
139
+ | 1.7853 | 0.3636 | 2 | 1.7201 |
140
+ | 1.3319 | 0.7273 | 4 | 1.3265 |
 
141
 
142
 
143
  ### Framework versions
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a047e9eca9c4d8ec6148b682b3006ca5e11f8e228de0d4b70abbc2b1f69d1d35
3
- size 432357050
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9db1f0813d4e6d03df3e4b97f45d791865f9b6c9f0a6ab2fca900a1dce5b7ae8
3
+ size 323103018