Simranjit commited on
Commit
0a4aa40
·
verified ·
1 Parent(s): 20efb2b

Training in progress, epoch 1

Browse files
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "models/phi-3-mini-128k-instruct",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -22,8 +22,8 @@
22
  "rank_pattern": {},
23
  "revision": null,
24
  "target_modules": [
25
- "o_proj",
26
- "qkv_proj"
27
  ],
28
  "task_type": "SEQ_CLASSIFICATION",
29
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "microsoft/Phi-3-mini-128k-instruct",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
22
  "rank_pattern": {},
23
  "revision": null,
24
  "target_modules": [
25
+ "qkv_proj",
26
+ "o_proj"
27
  ],
28
  "task_type": "SEQ_CLASSIFICATION",
29
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f3b8f0dea9eef8c60418af0f35a3e29b7acc4dc76f794b5b5814c505906e82a
3
- size 18916168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccff1fa6c478364af95295064013f402a9ee242a16c23b35e2cf31e47e9d4991
3
+ size 18965320
tokenizer.json CHANGED
@@ -1,11 +1,6 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 2048,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
  "padding": null,
10
  "added_tokens": [
11
  {
@@ -155,12 +150,6 @@
155
  "post_processor": {
156
  "type": "TemplateProcessing",
157
  "single": [
158
- {
159
- "SpecialToken": {
160
- "id": "<s>",
161
- "type_id": 0
162
- }
163
- },
164
  {
165
  "Sequence": {
166
  "id": "A",
@@ -169,24 +158,12 @@
169
  }
170
  ],
171
  "pair": [
172
- {
173
- "SpecialToken": {
174
- "id": "<s>",
175
- "type_id": 0
176
- }
177
- },
178
  {
179
  "Sequence": {
180
  "id": "A",
181
  "type_id": 0
182
  }
183
  },
184
- {
185
- "SpecialToken": {
186
- "id": "<s>",
187
- "type_id": 1
188
- }
189
- },
190
  {
191
  "Sequence": {
192
  "id": "B",
@@ -194,17 +171,7 @@
194
  }
195
  }
196
  ],
197
- "special_tokens": {
198
- "<s>": {
199
- "id": "<s>",
200
- "ids": [
201
- 1
202
- ],
203
- "tokens": [
204
- "<s>"
205
- ]
206
- }
207
- }
208
  },
209
  "decoder": {
210
  "type": "Sequence",
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
 
150
  "post_processor": {
151
  "type": "TemplateProcessing",
152
  "single": [
 
 
 
 
 
 
153
  {
154
  "Sequence": {
155
  "id": "A",
 
158
  }
159
  ],
160
  "pair": [
 
 
 
 
 
 
161
  {
162
  "Sequence": {
163
  "id": "A",
164
  "type_id": 0
165
  }
166
  },
 
 
 
 
 
 
167
  {
168
  "Sequence": {
169
  "id": "B",
 
171
  }
172
  }
173
  ],
174
+ "special_tokens": {}
 
 
 
 
 
 
 
 
 
 
175
  },
176
  "decoder": {
177
  "type": "Sequence",
tokenizer_config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "add_bos_token": true,
3
  "add_eos_token": false,
4
  "add_prefix_space": null,
5
  "added_tokens_decoder": {
@@ -117,7 +117,7 @@
117
  }
118
  },
119
  "bos_token": "<s>",
120
- "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') %}{{'<|user|>' + '\n' + message['content'] + '<|end|>' + '\n' + '<|assistant|>' + '\n'}}{% elif (message['role'] == 'assistant') %}{{message['content'] + '<|end|>' + '\n'}}{% endif %}{% endfor %}",
121
  "clean_up_tokenization_spaces": false,
122
  "eos_token": "<|endoftext|>",
123
  "legacy": false,
 
1
  {
2
+ "add_bos_token": false,
3
  "add_eos_token": false,
4
  "add_prefix_space": null,
5
  "added_tokens_decoder": {
 
117
  }
118
  },
119
  "bos_token": "<s>",
120
+ "chat_template": "{% for message in messages %}{% if message['role'] == 'system' %}{{'<|system|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}",
121
  "clean_up_tokenization_spaces": false,
122
  "eos_token": "<|endoftext|>",
123
  "legacy": false,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b333f43475f164649a6c47f303fe990edb8ba1922c23f23b7663c20aa57d1952
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a2de8d932e0f2a109536714b25257659595a5dcd467de13d6c0c868c47b1d08
3
  size 5176