Swephoenix commited on
Commit
53f1c21
·
verified ·
1 Parent(s): 880c896

Training in progress, step 50

Browse files
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "HuggingFaceTB/SmolLM2-1.7B-Instruct",
5
  "bias": "none",
6
  "eva_config": null,
7
  "exclude_modules": null,
@@ -23,9 +23,9 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
26
  "v_proj",
27
  "k_proj",
28
- "q_proj",
29
  "o_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "HuggingFaceTB/SmolLM-135M",
5
  "bias": "none",
6
  "eva_config": null,
7
  "exclude_modules": null,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "q_proj",
27
  "v_proj",
28
  "k_proj",
 
29
  "o_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3287ecf3f605e3b870ce6959ecd7da17cab4e6c879d7b5ecf469b13831d4cbaf
3
- size 6316688
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4a0fc7d6fd2fcbe03e54838b8829e7fa311b60eda112a3c45b3381a1201c14e
3
+ size 1874312
special_tokens_map.json CHANGED
@@ -1,29 +1,38 @@
1
  {
2
  "additional_special_tokens": [
 
3
  "<|im_start|>",
4
- "<|im_end|>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  ],
6
  "bos_token": {
7
- "content": "<|im_start|>",
8
  "lstrip": false,
9
  "normalized": false,
10
  "rstrip": false,
11
  "single_word": false
12
  },
13
  "eos_token": {
14
- "content": "<|im_end|>",
15
- "lstrip": false,
16
- "normalized": false,
17
- "rstrip": false,
18
- "single_word": false
19
- },
20
- "pad_token": {
21
- "content": "<|im_end|>",
22
  "lstrip": false,
23
  "normalized": false,
24
  "rstrip": false,
25
  "single_word": false
26
  },
 
27
  "unk_token": {
28
  "content": "<|endoftext|>",
29
  "lstrip": false,
 
1
  {
2
  "additional_special_tokens": [
3
+ "<|endoftext|>",
4
  "<|im_start|>",
5
+ "<|im_end|>",
6
+ "<repo_name>",
7
+ "<reponame>",
8
+ "<file_sep>",
9
+ "<filename>",
10
+ "<gh_stars>",
11
+ "<issue_start>",
12
+ "<issue_comment>",
13
+ "<issue_closed>",
14
+ "<jupyter_start>",
15
+ "<jupyter_text>",
16
+ "<jupyter_code>",
17
+ "<jupyter_output>",
18
+ "<jupyter_script>",
19
+ "<empty_output>"
20
  ],
21
  "bos_token": {
22
+ "content": "<|endoftext|>",
23
  "lstrip": false,
24
  "normalized": false,
25
  "rstrip": false,
26
  "single_word": false
27
  },
28
  "eos_token": {
29
+ "content": "<|endoftext|>",
 
 
 
 
 
 
 
30
  "lstrip": false,
31
  "normalized": false,
32
  "rstrip": false,
33
  "single_word": false
34
  },
35
+ "pad_token": "<|endoftext|>",
36
  "unk_token": {
37
  "content": "<|endoftext|>",
38
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -139,16 +139,31 @@
139
  }
140
  },
141
  "additional_special_tokens": [
 
142
  "<|im_start|>",
143
- "<|im_end|>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
144
  ],
145
- "bos_token": "<|im_start|>",
146
- "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful AI assistant named SmolLM, trained by Hugging Face<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
147
  "clean_up_tokenization_spaces": false,
148
- "eos_token": "<|im_end|>",
149
  "extra_special_tokens": {},
150
- "model_max_length": 8192,
151
- "pad_token": "<|im_end|>",
152
  "tokenizer_class": "GPT2Tokenizer",
153
  "unk_token": "<|endoftext|>",
154
  "vocab_size": 49152
 
139
  }
140
  },
141
  "additional_special_tokens": [
142
+ "<|endoftext|>",
143
  "<|im_start|>",
144
+ "<|im_end|>",
145
+ "<repo_name>",
146
+ "<reponame>",
147
+ "<file_sep>",
148
+ "<filename>",
149
+ "<gh_stars>",
150
+ "<issue_start>",
151
+ "<issue_comment>",
152
+ "<issue_closed>",
153
+ "<jupyter_start>",
154
+ "<jupyter_text>",
155
+ "<jupyter_code>",
156
+ "<jupyter_output>",
157
+ "<jupyter_script>",
158
+ "<empty_output>"
159
  ],
160
+ "bos_token": "<|endoftext|>",
161
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
162
  "clean_up_tokenization_spaces": false,
163
+ "eos_token": "<|endoftext|>",
164
  "extra_special_tokens": {},
165
+ "model_max_length": 1000000000000000019884624838656,
166
+ "pad_token": "<|endoftext|>",
167
  "tokenizer_class": "GPT2Tokenizer",
168
  "unk_token": "<|endoftext|>",
169
  "vocab_size": 49152
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7b85bfbc70a0765b49083e10772de4aa6ec753e20bfdb9c80e61c09428b5eb3a
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88660d09dd7b4ebf18d9049c01b24335b86e82962b0a55c970857804e9c824d4
3
  size 5304