danield12 commited on
Commit
053b775
·
verified ·
1 Parent(s): 4ae658e

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+
7
+ The following `bitsandbytes` quantization config was used during training:
8
+ - quant_method: bitsandbytes
9
+ - _load_in_8bit: False
10
+ - _load_in_4bit: True
11
+ - llm_int8_threshold: 6.0
12
+ - llm_int8_skip_modules: None
13
+ - llm_int8_enable_fp32_cpu_offload: False
14
+ - llm_int8_has_fp16_weight: False
15
+ - bnb_4bit_quant_type: nf4
16
+ - bnb_4bit_use_double_quant: True
17
+ - bnb_4bit_compute_dtype: bfloat16
18
+ - load_in_4bit: True
19
+ - load_in_8bit: False
20
+ ### Framework versions
21
+
22
+
23
+ - PEFT 0.4.0
adapter_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "meta-llama/Llama-2-7b-hf",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 64,
11
+ "lora_dropout": 0.1,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 16,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "up_proj",
18
+ "gate_proj",
19
+ "v_proj",
20
+ "q_proj",
21
+ "o_proj",
22
+ "down_proj",
23
+ "k_proj"
24
+ ],
25
+ "task_type": "SEQ_CLS"
26
+ }
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44dca49c744beb5b5154de61adad3fa409959e53d3efe0b94d4b6b4be819c8c7
3
+ size 160102926
all_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "total_flos": 1.0505428534296576e+17,
4
+ "train_loss": 0.5505467119013391,
5
+ "train_runtime": 36020.1073,
6
+ "train_samples_per_second": 0.146,
7
+ "train_steps_per_second": 0.002
8
+ }
runs/Mar19_18-58-37_11dfe7fa700f/events.out.tfevents.1710874719.11dfe7fa700f.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4e3a4eff3089375f416dc06bf51bacc2cc57bc459126a71f61ebf43c9de370a
3
+ size 12937
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "total_flos": 1.0505428534296576e+17,
4
+ "train_loss": 0.5505467119013391,
5
+ "train_runtime": 36020.1073,
6
+ "train_samples_per_second": 0.146,
7
+ "train_steps_per_second": 0.002
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,230 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.9969604863221885,
5
+ "eval_steps": 8,
6
+ "global_step": 82,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.1,
13
+ "grad_norm": 12.02409839630127,
14
+ "learning_rate": 1.8500000000000002e-05,
15
+ "loss": 1.0056,
16
+ "step": 8
17
+ },
18
+ {
19
+ "epoch": 0.1,
20
+ "eval_accuracy": 0.76,
21
+ "eval_f1_score": 0.76,
22
+ "eval_gmean": 0.7536682052115493,
23
+ "eval_loss": 0.6476712226867676,
24
+ "eval_precision": 0.76,
25
+ "eval_recall": 0.76,
26
+ "eval_runtime": 211.0576,
27
+ "eval_samples_per_second": 0.474,
28
+ "eval_steps_per_second": 0.062,
29
+ "step": 8
30
+ },
31
+ {
32
+ "epoch": 0.19,
33
+ "grad_norm": 13.794071197509766,
34
+ "learning_rate": 1.65e-05,
35
+ "loss": 0.761,
36
+ "step": 16
37
+ },
38
+ {
39
+ "epoch": 0.19,
40
+ "eval_accuracy": 0.8,
41
+ "eval_f1_score": 0.8,
42
+ "eval_gmean": 0.7947281139113244,
43
+ "eval_loss": 0.5526618361473083,
44
+ "eval_precision": 0.8,
45
+ "eval_recall": 0.8,
46
+ "eval_runtime": 211.7038,
47
+ "eval_samples_per_second": 0.472,
48
+ "eval_steps_per_second": 0.061,
49
+ "step": 16
50
+ },
51
+ {
52
+ "epoch": 0.29,
53
+ "grad_norm": 9.750018119812012,
54
+ "learning_rate": 1.45e-05,
55
+ "loss": 0.638,
56
+ "step": 24
57
+ },
58
+ {
59
+ "epoch": 0.29,
60
+ "eval_accuracy": 0.81,
61
+ "eval_f1_score": 0.8102915601023019,
62
+ "eval_gmean": 0.8066432910616553,
63
+ "eval_loss": 0.4845515191555023,
64
+ "eval_precision": 0.8107466340269278,
65
+ "eval_recall": 0.81,
66
+ "eval_runtime": 211.1147,
67
+ "eval_samples_per_second": 0.474,
68
+ "eval_steps_per_second": 0.062,
69
+ "step": 24
70
+ },
71
+ {
72
+ "epoch": 0.39,
73
+ "grad_norm": 9.329747200012207,
74
+ "learning_rate": 1.25e-05,
75
+ "loss": 0.6367,
76
+ "step": 32
77
+ },
78
+ {
79
+ "epoch": 0.39,
80
+ "eval_accuracy": 0.81,
81
+ "eval_f1_score": 0.8102915601023019,
82
+ "eval_gmean": 0.8066432910616553,
83
+ "eval_loss": 0.4284454584121704,
84
+ "eval_precision": 0.8107466340269278,
85
+ "eval_recall": 0.81,
86
+ "eval_runtime": 210.7064,
87
+ "eval_samples_per_second": 0.475,
88
+ "eval_steps_per_second": 0.062,
89
+ "step": 32
90
+ },
91
+ {
92
+ "epoch": 0.49,
93
+ "grad_norm": 9.867768287658691,
94
+ "learning_rate": 1.0500000000000001e-05,
95
+ "loss": 0.5932,
96
+ "step": 40
97
+ },
98
+ {
99
+ "epoch": 0.49,
100
+ "eval_accuracy": 0.82,
101
+ "eval_f1_score": 0.82,
102
+ "eval_gmean": 0.8152572201307815,
103
+ "eval_loss": 0.3858851194381714,
104
+ "eval_precision": 0.82,
105
+ "eval_recall": 0.82,
106
+ "eval_runtime": 211.0479,
107
+ "eval_samples_per_second": 0.474,
108
+ "eval_steps_per_second": 0.062,
109
+ "step": 40
110
+ },
111
+ {
112
+ "epoch": 0.58,
113
+ "grad_norm": 10.276468276977539,
114
+ "learning_rate": 8.5e-06,
115
+ "loss": 0.4603,
116
+ "step": 48
117
+ },
118
+ {
119
+ "epoch": 0.58,
120
+ "eval_accuracy": 0.83,
121
+ "eval_f1_score": 0.8297023993409536,
122
+ "eval_gmean": 0.823868981405519,
123
+ "eval_loss": 0.3537224233150482,
124
+ "eval_precision": 0.8295742042166185,
125
+ "eval_recall": 0.83,
126
+ "eval_runtime": 212.1782,
127
+ "eval_samples_per_second": 0.471,
128
+ "eval_steps_per_second": 0.061,
129
+ "step": 48
130
+ },
131
+ {
132
+ "epoch": 0.68,
133
+ "grad_norm": 8.40848445892334,
134
+ "learning_rate": 6.5000000000000004e-06,
135
+ "loss": 0.3794,
136
+ "step": 56
137
+ },
138
+ {
139
+ "epoch": 0.68,
140
+ "eval_accuracy": 0.86,
141
+ "eval_f1_score": 0.86,
142
+ "eval_gmean": 0.8563139396380225,
143
+ "eval_loss": 0.3282211720943451,
144
+ "eval_precision": 0.86,
145
+ "eval_recall": 0.86,
146
+ "eval_runtime": 211.0187,
147
+ "eval_samples_per_second": 0.474,
148
+ "eval_steps_per_second": 0.062,
149
+ "step": 56
150
+ },
151
+ {
152
+ "epoch": 0.78,
153
+ "grad_norm": 8.91843032836914,
154
+ "learning_rate": 4.5e-06,
155
+ "loss": 0.4213,
156
+ "step": 64
157
+ },
158
+ {
159
+ "epoch": 0.78,
160
+ "eval_accuracy": 0.87,
161
+ "eval_f1_score": 0.8701994884910484,
162
+ "eval_gmean": 0.8682247898524239,
163
+ "eval_loss": 0.31104254722595215,
164
+ "eval_precision": 0.8705752753977967,
165
+ "eval_recall": 0.87,
166
+ "eval_runtime": 211.0456,
167
+ "eval_samples_per_second": 0.474,
168
+ "eval_steps_per_second": 0.062,
169
+ "step": 64
170
+ },
171
+ {
172
+ "epoch": 0.88,
173
+ "grad_norm": 6.700900554656982,
174
+ "learning_rate": 2.5e-06,
175
+ "loss": 0.3611,
176
+ "step": 72
177
+ },
178
+ {
179
+ "epoch": 0.88,
180
+ "eval_accuracy": 0.88,
181
+ "eval_f1_score": 0.8803427172582619,
182
+ "eval_gmean": 0.8801313530867828,
183
+ "eval_loss": 0.30110642313957214,
184
+ "eval_precision": 0.881396103896104,
185
+ "eval_recall": 0.88,
186
+ "eval_runtime": 211.6438,
187
+ "eval_samples_per_second": 0.472,
188
+ "eval_steps_per_second": 0.061,
189
+ "step": 72
190
+ },
191
+ {
192
+ "epoch": 0.97,
193
+ "grad_norm": 7.661783695220947,
194
+ "learning_rate": 5.000000000000001e-07,
195
+ "loss": 0.3248,
196
+ "step": 80
197
+ },
198
+ {
199
+ "epoch": 0.97,
200
+ "eval_accuracy": 0.88,
201
+ "eval_f1_score": 0.8803427172582619,
202
+ "eval_gmean": 0.8801313530867828,
203
+ "eval_loss": 0.29578596353530884,
204
+ "eval_precision": 0.881396103896104,
205
+ "eval_recall": 0.88,
206
+ "eval_runtime": 210.1348,
207
+ "eval_samples_per_second": 0.476,
208
+ "eval_steps_per_second": 0.062,
209
+ "step": 80
210
+ },
211
+ {
212
+ "epoch": 1.0,
213
+ "step": 82,
214
+ "total_flos": 1.0505428534296576e+17,
215
+ "train_loss": 0.5505467119013391,
216
+ "train_runtime": 36020.1073,
217
+ "train_samples_per_second": 0.146,
218
+ "train_steps_per_second": 0.002
219
+ }
220
+ ],
221
+ "logging_steps": 8,
222
+ "max_steps": 82,
223
+ "num_input_tokens_seen": 0,
224
+ "num_train_epochs": 1,
225
+ "save_steps": 500,
226
+ "total_flos": 1.0505428534296576e+17,
227
+ "train_batch_size": 16,
228
+ "trial_name": null,
229
+ "trial_params": null
230
+ }