kurtn718 commited on
Commit
7a3db01
·
1 Parent(s): 06a943e

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+ - PEFT 0.5.0
9
+
10
+ - PEFT 0.5.0
adapter_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "tiiuae/falcon-7b",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 16,
11
+ "lora_dropout": 0.1,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 32,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "query_key_value",
18
+ "dense",
19
+ "dense_h_to_4h",
20
+ "dense_4h_to_h"
21
+ ],
22
+ "task_type": "CAUSAL_LM"
23
+ }
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9346b9bdb0b0d7fc3e555dc21f401422f73a7102e7e1c5fa849cc95fa764ef5a
3
+ size 261189453
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d56167de0cbb796019c6b1df86d9c1122876c4cae810d8231c8a5b6a9adb6f4
3
+ size 522348805
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bfc96dc9be5642d06c809572ec989df6b2ec19869f171677758e9ae9d5fa071
3
+ size 14575
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dc9f914380ce6a3c801c27f7620226b1ff060a4b0603d5fa3205fe837822f49
3
+ size 627
special_tokens_map.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ ">>TITLE<<",
4
+ ">>ABSTRACT<<",
5
+ ">>INTRODUCTION<<",
6
+ ">>SUMMARY<<",
7
+ ">>COMMENT<<",
8
+ ">>ANSWER<<",
9
+ ">>QUESTION<<",
10
+ ">>DOMAIN<<",
11
+ ">>PREFIX<<",
12
+ ">>SUFFIX<<",
13
+ ">>MIDDLE<<"
14
+ ],
15
+ "eos_token": "<|endoftext|>",
16
+ "pad_token": "<|endoftext|>"
17
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,122 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": ">>TITLE<<",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": ">>ABSTRACT<<",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": ">>INTRODUCTION<<",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": ">>SUMMARY<<",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "4": {
37
+ "content": ">>COMMENT<<",
38
+ "lstrip": false,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ },
44
+ "5": {
45
+ "content": ">>ANSWER<<",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false,
50
+ "special": true
51
+ },
52
+ "6": {
53
+ "content": ">>QUESTION<<",
54
+ "lstrip": false,
55
+ "normalized": false,
56
+ "rstrip": false,
57
+ "single_word": false,
58
+ "special": true
59
+ },
60
+ "7": {
61
+ "content": ">>DOMAIN<<",
62
+ "lstrip": false,
63
+ "normalized": false,
64
+ "rstrip": false,
65
+ "single_word": false,
66
+ "special": true
67
+ },
68
+ "8": {
69
+ "content": ">>PREFIX<<",
70
+ "lstrip": false,
71
+ "normalized": false,
72
+ "rstrip": false,
73
+ "single_word": false,
74
+ "special": true
75
+ },
76
+ "9": {
77
+ "content": ">>SUFFIX<<",
78
+ "lstrip": false,
79
+ "normalized": false,
80
+ "rstrip": false,
81
+ "single_word": false,
82
+ "special": true
83
+ },
84
+ "10": {
85
+ "content": ">>MIDDLE<<",
86
+ "lstrip": false,
87
+ "normalized": false,
88
+ "rstrip": false,
89
+ "single_word": false,
90
+ "special": true
91
+ },
92
+ "11": {
93
+ "content": "<|endoftext|>",
94
+ "lstrip": false,
95
+ "normalized": false,
96
+ "rstrip": false,
97
+ "single_word": false,
98
+ "special": true
99
+ }
100
+ },
101
+ "additional_special_tokens": [
102
+ ">>TITLE<<",
103
+ ">>ABSTRACT<<",
104
+ ">>INTRODUCTION<<",
105
+ ">>SUMMARY<<",
106
+ ">>COMMENT<<",
107
+ ">>ANSWER<<",
108
+ ">>QUESTION<<",
109
+ ">>DOMAIN<<",
110
+ ">>PREFIX<<",
111
+ ">>SUFFIX<<",
112
+ ">>MIDDLE<<"
113
+ ],
114
+ "clean_up_tokenization_spaces": true,
115
+ "eos_token": "<|endoftext|>",
116
+ "model_input_names": [
117
+ "input_ids",
118
+ "attention_mask"
119
+ ],
120
+ "model_max_length": 2048,
121
+ "tokenizer_class": "PreTrainedTokenizerFast"
122
+ }
trainer_state.json ADDED
@@ -0,0 +1,179 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 1.4523780345916748,
3
+ "best_model_checkpoint": "./results/checkpoint-64",
4
+ "epoch": 0.05200081251269551,
5
+ "eval_steps": 8,
6
+ "global_step": 64,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.0,
13
+ "learning_rate": 0.0002,
14
+ "loss": 1.604,
15
+ "step": 4
16
+ },
17
+ {
18
+ "epoch": 0.01,
19
+ "learning_rate": 0.0002,
20
+ "loss": 1.5865,
21
+ "step": 8
22
+ },
23
+ {
24
+ "epoch": 0.01,
25
+ "eval_loss": 1.58369779586792,
26
+ "eval_runtime": 33.9799,
27
+ "eval_samples_per_second": 15.244,
28
+ "eval_steps_per_second": 1.913,
29
+ "step": 8
30
+ },
31
+ {
32
+ "epoch": 0.01,
33
+ "learning_rate": 0.0002,
34
+ "loss": 1.4024,
35
+ "step": 12
36
+ },
37
+ {
38
+ "epoch": 0.01,
39
+ "learning_rate": 0.0002,
40
+ "loss": 1.3627,
41
+ "step": 16
42
+ },
43
+ {
44
+ "epoch": 0.01,
45
+ "eval_loss": 1.503720998764038,
46
+ "eval_runtime": 34.2342,
47
+ "eval_samples_per_second": 15.131,
48
+ "eval_steps_per_second": 1.899,
49
+ "step": 16
50
+ },
51
+ {
52
+ "epoch": 0.02,
53
+ "learning_rate": 0.0002,
54
+ "loss": 1.3242,
55
+ "step": 20
56
+ },
57
+ {
58
+ "epoch": 0.02,
59
+ "learning_rate": 0.0002,
60
+ "loss": 1.2733,
61
+ "step": 24
62
+ },
63
+ {
64
+ "epoch": 0.02,
65
+ "eval_loss": 1.4826295375823975,
66
+ "eval_runtime": 34.4234,
67
+ "eval_samples_per_second": 15.048,
68
+ "eval_steps_per_second": 1.888,
69
+ "step": 24
70
+ },
71
+ {
72
+ "epoch": 0.02,
73
+ "learning_rate": 0.0002,
74
+ "loss": 1.507,
75
+ "step": 28
76
+ },
77
+ {
78
+ "epoch": 0.03,
79
+ "learning_rate": 0.0002,
80
+ "loss": 1.472,
81
+ "step": 32
82
+ },
83
+ {
84
+ "epoch": 0.03,
85
+ "eval_loss": 1.4713466167449951,
86
+ "eval_runtime": 34.5227,
87
+ "eval_samples_per_second": 15.005,
88
+ "eval_steps_per_second": 1.883,
89
+ "step": 32
90
+ },
91
+ {
92
+ "epoch": 0.03,
93
+ "learning_rate": 0.0002,
94
+ "loss": 1.4269,
95
+ "step": 36
96
+ },
97
+ {
98
+ "epoch": 0.03,
99
+ "learning_rate": 0.0002,
100
+ "loss": 1.6282,
101
+ "step": 40
102
+ },
103
+ {
104
+ "epoch": 0.03,
105
+ "eval_loss": 1.4705039262771606,
106
+ "eval_runtime": 34.5472,
107
+ "eval_samples_per_second": 14.994,
108
+ "eval_steps_per_second": 1.881,
109
+ "step": 40
110
+ },
111
+ {
112
+ "epoch": 0.04,
113
+ "learning_rate": 0.0002,
114
+ "loss": 1.6744,
115
+ "step": 44
116
+ },
117
+ {
118
+ "epoch": 0.04,
119
+ "learning_rate": 0.0002,
120
+ "loss": 1.789,
121
+ "step": 48
122
+ },
123
+ {
124
+ "epoch": 0.04,
125
+ "eval_loss": 1.4801452159881592,
126
+ "eval_runtime": 34.5611,
127
+ "eval_samples_per_second": 14.988,
128
+ "eval_steps_per_second": 1.881,
129
+ "step": 48
130
+ },
131
+ {
132
+ "epoch": 0.04,
133
+ "learning_rate": 0.0002,
134
+ "loss": 1.7547,
135
+ "step": 52
136
+ },
137
+ {
138
+ "epoch": 0.05,
139
+ "learning_rate": 0.0002,
140
+ "loss": 1.3081,
141
+ "step": 56
142
+ },
143
+ {
144
+ "epoch": 0.05,
145
+ "eval_loss": 1.465359091758728,
146
+ "eval_runtime": 34.5837,
147
+ "eval_samples_per_second": 14.978,
148
+ "eval_steps_per_second": 1.88,
149
+ "step": 56
150
+ },
151
+ {
152
+ "epoch": 0.05,
153
+ "learning_rate": 0.0002,
154
+ "loss": 1.2313,
155
+ "step": 60
156
+ },
157
+ {
158
+ "epoch": 0.05,
159
+ "learning_rate": 0.0002,
160
+ "loss": 1.4273,
161
+ "step": 64
162
+ },
163
+ {
164
+ "epoch": 0.05,
165
+ "eval_loss": 1.4523780345916748,
166
+ "eval_runtime": 34.6395,
167
+ "eval_samples_per_second": 14.954,
168
+ "eval_steps_per_second": 1.876,
169
+ "step": 64
170
+ }
171
+ ],
172
+ "logging_steps": 4,
173
+ "max_steps": 64,
174
+ "num_train_epochs": 1,
175
+ "save_steps": 8,
176
+ "total_flos": 7578875318034432.0,
177
+ "trial_name": null,
178
+ "trial_params": null
179
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c344a89cdc0be10f4f9108a8b6a79229619c02326bc5d60fb8aa724a57e46f6b
3
+ size 4027