samoline commited on
Commit
8a3741d
·
verified ·
1 Parent(s): 1c52540

End of training

Browse files
Files changed (2) hide show
  1. README.md +12 -9
  2. adapter_model.bin +1 -1
README.md CHANGED
@@ -24,12 +24,15 @@ bf16: auto
24
  chat_template: llama3
25
  dataset_prepared_path: null
26
  datasets:
27
- - format: custom
28
- path: mhenrichsen/alpaca_2k_test
 
 
 
29
  type:
30
- field_input: input
31
- field_instruction: instruction
32
- field_output: output
33
  format: '{instruction} {input}'
34
  no_input_format: '{instruction}'
35
  system_format: '{system}'
@@ -65,7 +68,7 @@ lora_target_linear: true
65
  lr_scheduler: cosine
66
  max_steps: 2
67
  micro_batch_size: 1
68
- mlflow_experiment_name: mhenrichsen/alpaca_2k_test
69
  model_type: AutoModelForCausalLM
70
  num_epochs: 1
71
  optimizer: adamw_bnb_8bit
@@ -102,7 +105,7 @@ xformers_attention: null
102
 
103
  This model is a fine-tuned version of [Maykeye/TinyLLama-v0](https://huggingface.co/Maykeye/TinyLLama-v0) on the None dataset.
104
  It achieves the following results on the evaluation set:
105
- - Loss: 7.2158
106
 
107
  ## Model description
108
 
@@ -134,8 +137,8 @@ The following hyperparameters were used during training:
134
 
135
  | Training Loss | Epoch | Step | Validation Loss |
136
  |:-------------:|:------:|:----:|:---------------:|
137
- | 7.6993 | 0.0005 | 1 | 7.2157 |
138
- | 8.0312 | 0.0011 | 2 | 7.2158 |
139
 
140
 
141
  ### Framework versions
 
24
  chat_template: llama3
25
  dataset_prepared_path: null
26
  datasets:
27
+ - data_files:
28
+ - orca_dpo_pairs.json
29
+ ds_type: json
30
+ format: custom
31
+ path: /workspace/input_data/orca_dpo_pairs.json
32
  type:
33
+ field_input: system
34
+ field_instruction: question
35
+ field_output: chosen
36
  format: '{instruction} {input}'
37
  no_input_format: '{instruction}'
38
  system_format: '{system}'
 
68
  lr_scheduler: cosine
69
  max_steps: 2
70
  micro_batch_size: 1
71
+ mlflow_experiment_name: /tmp/orca_dpo_pairs.json
72
  model_type: AutoModelForCausalLM
73
  num_epochs: 1
74
  optimizer: adamw_bnb_8bit
 
105
 
106
  This model is a fine-tuned version of [Maykeye/TinyLLama-v0](https://huggingface.co/Maykeye/TinyLLama-v0) on the None dataset.
107
  It achieves the following results on the evaluation set:
108
+ - Loss: 7.8181
109
 
110
  ## Model description
111
 
 
137
 
138
  | Training Loss | Epoch | Step | Validation Loss |
139
  |:-------------:|:------:|:----:|:---------------:|
140
+ | 6.0547 | 0.0001 | 1 | 7.8181 |
141
+ | 7.7655 | 0.0002 | 2 | 7.8181 |
142
 
143
 
144
  ### Framework versions
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7c2ac9278d5675b2475deb883e9e6637fa15b3e4177d55b5bd6d2b18a3513287
3
  size 228490
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f883ed1767b612705d4b8bf50290aba225a4a968847708521c890be5646cff2e
3
  size 228490