Incomple commited on
Commit
e5abe0b
·
verified ·
1 Parent(s): 389baf5

End of training

Browse files
Files changed (4) hide show
  1. README.md +3 -3
  2. all_results.json +3 -3
  3. train_results.json +3 -3
  4. trainer_state.json +3 -3
README.md CHANGED
@@ -3,10 +3,10 @@ library_name: peft
3
  license: llama3.1
4
  base_model: meta-llama/Llama-3.1-8B-Instruct
5
  tags:
6
- - trl
7
- - dpo
8
  - llama-factory
9
  - lora
 
 
10
  - generated_from_trainer
11
  model-index:
12
  - name: Llama-3.1-8B-Instruct_baseline
@@ -18,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # Llama-3.1-8B-Instruct_baseline
20
 
21
- This model is a fine-tuned version of [meta-llama/Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct) on an unknown dataset.
22
 
23
  ## Model description
24
 
 
3
  license: llama3.1
4
  base_model: meta-llama/Llama-3.1-8B-Instruct
5
  tags:
 
 
6
  - llama-factory
7
  - lora
8
+ - trl
9
+ - dpo
10
  - generated_from_trainer
11
  model-index:
12
  - name: Llama-3.1-8B-Instruct_baseline
 
18
 
19
  # Llama-3.1-8B-Instruct_baseline
20
 
21
+ This model is a fine-tuned version of [meta-llama/Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct) on the baseline dataset.
22
 
23
  ## Model description
24
 
all_results.json CHANGED
@@ -2,7 +2,7 @@
2
  "epoch": 0.9942196531791907,
3
  "total_flos": 8626483251118080.0,
4
  "train_loss": 0.6937350367390832,
5
- "train_runtime": 232.7569,
6
- "train_samples_per_second": 1.482,
7
- "train_steps_per_second": 0.185
8
  }
 
2
  "epoch": 0.9942196531791907,
3
  "total_flos": 8626483251118080.0,
4
  "train_loss": 0.6937350367390832,
5
+ "train_runtime": 231.1609,
6
+ "train_samples_per_second": 1.492,
7
+ "train_steps_per_second": 0.186
8
  }
train_results.json CHANGED
@@ -2,7 +2,7 @@
2
  "epoch": 0.9942196531791907,
3
  "total_flos": 8626483251118080.0,
4
  "train_loss": 0.6937350367390832,
5
- "train_runtime": 232.7569,
6
- "train_samples_per_second": 1.482,
7
- "train_steps_per_second": 0.185
8
  }
 
2
  "epoch": 0.9942196531791907,
3
  "total_flos": 8626483251118080.0,
4
  "train_loss": 0.6937350367390832,
5
+ "train_runtime": 231.1609,
6
+ "train_samples_per_second": 1.492,
7
+ "train_steps_per_second": 0.186
8
  }
trainer_state.json CHANGED
@@ -223,9 +223,9 @@
223
  "step": 43,
224
  "total_flos": 8626483251118080.0,
225
  "train_loss": 0.6937350367390832,
226
- "train_runtime": 232.7569,
227
- "train_samples_per_second": 1.482,
228
- "train_steps_per_second": 0.185
229
  }
230
  ],
231
  "logging_steps": 3,
 
223
  "step": 43,
224
  "total_flos": 8626483251118080.0,
225
  "train_loss": 0.6937350367390832,
226
+ "train_runtime": 231.1609,
227
+ "train_samples_per_second": 1.492,
228
+ "train_steps_per_second": 0.186
229
  }
230
  ],
231
  "logging_steps": 3,