Heralax commited on
Commit
c1841b4
·
verified ·
1 Parent(s): 7d10e25

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -170
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  library_name: transformers
3
- license: apache-2.0
4
  base_model: Heralax/test-model-5-pretrain
5
  tags:
6
  - axolotl
@@ -13,181 +13,39 @@ datasets:
13
  - factual_sft_completion/combined_all_2.jsonl
14
  - factual_sft_completion/combined_all_3.jsonl
15
  - factual_sft_completion/combined_all_1.jsonl
16
- - generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_1081745.jsonl
17
- - generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_534422.jsonl
18
- - generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_1068845.jsonl
19
- - generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_1081745.jsonl
20
- - generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_1081745.jsonl
21
- - generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_534422.jsonl
 
 
 
 
 
 
22
  - generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_534422.jsonl
23
- - generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_4326980.jsonl
24
- - generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_1081745.jsonl
25
- - generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_2137691.jsonl
26
- - generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_534422.jsonl
27
- - generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_2163490.jsonl
 
 
 
 
 
28
  model-index:
29
  - name: test-model-5-sft
30
  results: []
31
  ---
32
 
33
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
34
- should probably proofread and complete it, then remove this comment. -->
35
 
36
- [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
37
- <details><summary>See axolotl config</summary>
38
-
39
- axolotl version: `0.10.0.dev0`
40
- ```yaml
41
- base_model: Heralax/test-model-5-pretrain
42
- tokenizer_type: AutoTokenizer
43
- model_type: AutoModelForCausalLM
44
- load_in_8bit: false
45
- load_in_4bit: false
46
- strict: false
47
- datasets:
48
- - path: axolotl_rag_conversations_facts.jsonl
49
- type: input_output
50
- - path: axolotl_correction_conversations_facts.json
51
- type: input_output
52
- - path: pretraining_subset_2170418.jsonl
53
- type: completion
54
- - path: factual_sft_completion/combined_all_0.jsonl
55
- type: completion
56
- - path: factual_sft_completion/combined_all_2.jsonl
57
- type: completion
58
- - path: factual_sft_completion/combined_all_3.jsonl
59
- type: completion
60
- - path: factual_sft_completion/combined_all_1.jsonl
61
- type: completion
62
- - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_1081745.jsonl
63
- type: completion
64
- - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_534422.jsonl
65
- type: completion
66
- - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_1068845.jsonl
67
- type: completion
68
- - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_1081745.jsonl
69
- type: completion
70
- - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_1081745.jsonl
71
- type: completion
72
- - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_534422.jsonl
73
- type: completion
74
- - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_534422.jsonl
75
- type: completion
76
- - path: generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_4326980.jsonl
77
- type: completion
78
- - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_1081745.jsonl
79
- type: completion
80
- - path: generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_2137691.jsonl
81
- type: completion
82
- - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_534422.jsonl
83
- type: completion
84
- - path: generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_2163490.jsonl
85
- type: completion
86
- dataset_prepared_path: last_finetune_prepared
87
- output_dir: ./finetune-model-output
88
- seed: 1337
89
- sequence_len: 5000
90
- sample_packing: true
91
- pad_to_sequence_len: false
92
- shuffle_merged_datasets: true
93
- gradient_accumulation_steps: 75
94
- micro_batch_size: 2
95
- eval_batch_size: 4
96
- num_epochs: 5
97
- optimizer: paged_adamw_8bit
98
- lr_scheduler: constant
99
- learning_rate: 2.0e-05
100
- noisy_embedding_alpha: 5
101
- weight_decay: 0
102
- train_on_inputs: false
103
- group_by_length: false
104
- bf16: true
105
- fp16: false
106
- tf32: false
107
- gradient_checkpointing: true
108
- logging_steps: 1
109
- xformers_attention: false
110
- flash_attention: true
111
- chat_template: chatml
112
- auto_resume_from_checkpoints: false
113
- warmup_ratio: 0.1
114
- evals_per_epoch: 1
115
- val_set_size: 0.04
116
- saves_per_epoch: 1
117
- eval_sample_packing: false
118
- save_total_limit: 2
119
- special_tokens:
120
- pad_token: <unk>
121
- use_liger_kernel: true
122
- plugins:
123
- - axolotl.integrations.liger.LigerPlugin
124
- liger_rope: true
125
- liger_rms_norm: true
126
- liger_glu_activation: true
127
- liger_layer_norm: true
128
- liger_fused_linear_cross_entropy: true
129
- sequence_length: 10000
130
- wandb_project: test-project
131
- wandb_entity: ''
132
- wandb_watch: ''
133
- wandb_run_id: ''
134
- wandb_log_model: ''
135
- hub_model_id: Heralax/test-model-5-sft
136
- hub_strategy: all_checkpoints
137
-
138
- ```
139
-
140
- </details><br>
141
-
142
- # test-model-5-sft
143
-
144
- This model is a fine-tuned version of [Heralax/test-model-5-pretrain](https://huggingface.co/Heralax/test-model-5-pretrain) on the axolotl_rag_conversations_facts.jsonl, the axolotl_correction_conversations_facts.json, the pretraining_subset_2170418.jsonl, the factual_sft_completion/combined_all_0.jsonl, the factual_sft_completion/combined_all_2.jsonl, the factual_sft_completion/combined_all_3.jsonl, the factual_sft_completion/combined_all_1.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_1081745.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_534422.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_1068845.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_1081745.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_1081745.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_534422.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_534422.jsonl, the generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_4326980.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_1081745.jsonl, the generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_2137691.jsonl, the generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_534422.jsonl and the generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_2163490.jsonl datasets.
145
- It achieves the following results on the evaluation set:
146
  - Loss: 0.6264
147
 
148
- ## Model description
149
-
150
- More information needed
151
-
152
- ## Intended uses & limitations
153
-
154
- More information needed
155
-
156
- ## Training and evaluation data
157
-
158
- More information needed
159
-
160
- ## Training procedure
161
-
162
- ### Training hyperparameters
163
-
164
- The following hyperparameters were used during training:
165
- - learning_rate: 2e-05
166
- - train_batch_size: 2
167
- - eval_batch_size: 4
168
- - seed: 1337
169
- - gradient_accumulation_steps: 75
170
- - total_train_batch_size: 150
171
- - optimizer: Use OptimizerNames.PAGED_ADAMW_8BIT with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
172
- - lr_scheduler_type: constant
173
- - lr_scheduler_warmup_steps: 20
174
- - training_steps: 205
175
-
176
- ### Training results
177
-
178
- | Training Loss | Epoch | Step | Validation Loss |
179
- |:-------------:|:------:|:----:|:---------------:|
180
- | 1.6475 | 0.0240 | 1 | 1.5248 |
181
- | 0.6333 | 0.9856 | 41 | 0.5850 |
182
- | 0.4419 | 1.9615 | 82 | 0.5704 |
183
- | 0.2823 | 2.9375 | 123 | 0.5763 |
184
- | 0.2005 | 3.9135 | 164 | 0.6002 |
185
- | 0.1387 | 4.8894 | 205 | 0.6264 |
186
-
187
-
188
- ### Framework versions
189
-
190
- - Transformers 4.52.3
191
- - Pytorch 2.6.0+cu124
192
- - Datasets 3.6.0
193
- - Tokenizers 0.21.1
 
1
  ---
2
  library_name: transformers
3
+ license: llama3.1
4
  base_model: Heralax/test-model-5-pretrain
5
  tags:
6
  - axolotl
 
13
  - factual_sft_completion/combined_all_2.jsonl
14
  - factual_sft_completion/combined_all_3.jsonl
15
  - factual_sft_completion/combined_all_1.jsonl
16
+ - >-
17
+ generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_1081745.jsonl
18
+ - >-
19
+ generic_sft_completion/Augmentoolkit-Augmentoolkit-LMsys-800k-Thoughts_534422.jsonl
20
+ - >-
21
+ generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_1068845.jsonl
22
+ - >-
23
+ generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_1081745.jsonl
24
+ - >-
25
+ generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_1081745.jsonl
26
+ - >-
27
+ generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_534422.jsonl
28
  - generic_sft_completion/Augmentoolkit-Augmentoolkit-Pippa-Thoughts_534422.jsonl
29
+ - >-
30
+ generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_4326980.jsonl
31
+ - >-
32
+ generic_sft_completion/Augmentoolkit-Augmentoolkit-Capybara-2point5mil-Thoughts_1081745.jsonl
33
+ - >-
34
+ generic_sft_completion/Augmentoolkit-Openthoughts-100mil-DifferentFormat_2137691.jsonl
35
+ - >-
36
+ generic_sft_completion/Augmentoolkit-Augmentoolkit-Bluemoon-1mil-thoughts_534422.jsonl
37
+ - >-
38
+ generic_sft_completion/Augmentoolkit-Augmentoolkit-Generic-Grabbag-Thoughts_2163490.jsonl
39
  model-index:
40
  - name: test-model-5-sft
41
  results: []
42
  ---
43
 
44
+ # llama-Augmentoolkit-MilitaryModel-Demo-NotUndertrained
 
45
 
46
+ This model achieves the following results on the evaluation set:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
  - Loss: 0.6264
48
 
49
+ This is a less-undertrained version of one of the demo factual models (the military one). Both such models were a bit undertrained. This one suffers from that less and should produce better results (theoretically, I have not tested it yet).
50
+ Same prompt as the military one.
51
+ Try this model out!