besimray commited on
Commit
38ca46c
·
verified ·
1 Parent(s): 8df8945

End of training

Browse files
Files changed (3) hide show
  1. README.md +12 -12
  2. adapter_model.bin +1 -1
  3. adapter_model.safetensors +1 -1
README.md CHANGED
@@ -30,8 +30,8 @@ datasets:
30
  path: /workspace/input_data/MATH-Hard_train_data.json
31
  type:
32
  field_input: problem
33
- field_instruction: type
34
- field_output: solution
35
  system_format: '{system}'
36
  system_prompt: ''
37
  debug: null
@@ -86,7 +86,7 @@ wandb_entity: besimray24-rayon
86
  wandb_mode: online
87
  wandb_project: Public_TuningSN
88
  wandb_run: miner_id_24
89
- wandb_runid: 3e895734-fc32-4a73-b997-346d471cdefc
90
  warmup_steps: 10
91
  weight_decay: 0.01
92
  xformers_attention: null
@@ -99,7 +99,7 @@ xformers_attention: null
99
 
100
  This model is a fine-tuned version of [unsloth/Llama-3.2-1B-Instruct](https://huggingface.co/unsloth/Llama-3.2-1B-Instruct) on the None dataset.
101
  It achieves the following results on the evaluation set:
102
- - Loss: 0.7634
103
 
104
  ## Model description
105
 
@@ -133,14 +133,14 @@ The following hyperparameters were used during training:
133
 
134
  | Training Loss | Epoch | Step | Validation Loss |
135
  |:-------------:|:------:|:----:|:---------------:|
136
- | 0.9595 | 0.0129 | 1 | 0.9746 |
137
- | 0.9389 | 0.2572 | 20 | 0.8177 |
138
- | 0.7865 | 0.5145 | 40 | 0.7905 |
139
- | 0.7565 | 0.7717 | 60 | 0.7795 |
140
- | 0.8137 | 1.0289 | 80 | 0.7711 |
141
- | 0.7243 | 1.2862 | 100 | 0.7675 |
142
- | 0.8195 | 1.5434 | 120 | 0.7645 |
143
- | 0.6793 | 1.8006 | 140 | 0.7634 |
144
 
145
 
146
  ### Framework versions
 
30
  path: /workspace/input_data/MATH-Hard_train_data.json
31
  type:
32
  field_input: problem
33
+ field_instruction: solution
34
+ field_output: type
35
  system_format: '{system}'
36
  system_prompt: ''
37
  debug: null
 
86
  wandb_mode: online
87
  wandb_project: Public_TuningSN
88
  wandb_run: miner_id_24
89
+ wandb_runid: 3882ca50-f5d8-4a62-83cf-33e7720e8c52
90
  warmup_steps: 10
91
  weight_decay: 0.01
92
  xformers_attention: null
 
99
 
100
  This model is a fine-tuned version of [unsloth/Llama-3.2-1B-Instruct](https://huggingface.co/unsloth/Llama-3.2-1B-Instruct) on the None dataset.
101
  It achieves the following results on the evaluation set:
102
+ - Loss: 0.0766
103
 
104
  ## Model description
105
 
 
133
 
134
  | Training Loss | Epoch | Step | Validation Loss |
135
  |:-------------:|:------:|:----:|:---------------:|
136
+ | 9.1 | 0.0129 | 1 | 8.9962 |
137
+ | 0.3746 | 0.2572 | 20 | 0.3471 |
138
+ | 0.2618 | 0.5145 | 40 | 0.1247 |
139
+ | 0.106 | 0.7717 | 60 | 0.1141 |
140
+ | 0.1457 | 1.0289 | 80 | 0.1035 |
141
+ | 0.0493 | 1.2862 | 100 | 0.0947 |
142
+ | 0.1237 | 1.5434 | 120 | 0.0765 |
143
+ | 0.0294 | 1.8006 | 140 | 0.0766 |
144
 
145
 
146
  ### Framework versions
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da147b15838823004a0f5347d410eb528606bf55c3fa906cd13390d08d5201dd
3
  size 45169354
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9e5432c0904d2ebf3acbe3d9dc2636391218bc21433147b98fca1a4e966c12f
3
  size 45169354
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1fa8d826125079ec9c5474ab65c06145915471c18bc66f1805c7d717cadd029e
3
  size 45118424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:833eb01138783f900186e10d4e4abdce2331f847c08833bd0fa504402085a60b
3
  size 45118424