ravi1289 commited on
Commit
65b6e3b
Β·
verified Β·
1 Parent(s): 0bd3042

Upload fine-tuned CodeMind model

Browse files
checkpoints/checkpoint-94/README.md CHANGED
@@ -1,9 +1,9 @@
1
  ---
2
- base_model: codellama/CodeLlama-7b-hf
3
  library_name: peft
4
  pipeline_tag: text-generation
5
  tags:
6
- - base_model:adapter:codellama/CodeLlama-7b-hf
7
  - lora
8
  - transformers
9
  ---
@@ -204,4 +204,5 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
204
  [More Information Needed]
205
  ### Framework versions
206
 
 
207
  - PEFT 0.17.0
 
1
  ---
2
+ base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
3
  library_name: peft
4
  pipeline_tag: text-generation
5
  tags:
6
+ - base_model:adapter:TinyLlama/TinyLlama-1.1B-Chat-v1.0
7
  - lora
8
  - transformers
9
  ---
 
204
  [More Information Needed]
205
  ### Framework versions
206
 
207
+ - PEFT 0.17.1
208
  - PEFT 0.17.0
checkpoints/checkpoint-94/adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "codellama/CodeLlama-7b-hf",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
checkpoints/checkpoint-94/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a6ad631c5d6c6d209ebd70a23c865e4b110c827017f205b18a2432ebbc554f3f
3
- size 16794200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8f402414502e7546231249c382c071c163a58eec09838b31328d00db17f99cf
3
+ size 4517152
checkpoints/checkpoint-94/chat_template.jinja ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {% for message in messages %}
2
+ {% if message['role'] == 'user' %}
3
+ {{ '<|user|>
4
+ ' + message['content'] + eos_token }}
5
+ {% elif message['role'] == 'system' %}
6
+ {{ '<|system|>
7
+ ' + message['content'] + eos_token }}
8
+ {% elif message['role'] == 'assistant' %}
9
+ {{ '<|assistant|>
10
+ ' + message['content'] + eos_token }}
11
+ {% endif %}
12
+ {% if loop.last and add_generation_prompt %}
13
+ {{ '<|assistant|>' }}
14
+ {% endif %}
15
+ {% endfor %}
checkpoints/checkpoint-94/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee59e677a224318171cc166414866e279b0dd120f4ca041953190e49dd4e0372
3
- size 33662074
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef61894ea730991e33f2cce894aeac277458a3e3daab7ae2de9a95f69606b3d2
3
+ size 9087051
checkpoints/checkpoint-94/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3614fc15a38e35eb840809f67f4861938ac6b348a9177a9d9742a1d9d167a7cf
3
- size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ebc6a66c38d82fae596405cd59f9b05aa195310c5e9d92a33a3e411244cc412
3
+ size 14645
checkpoints/checkpoint-94/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c727a1fa2ac82030a89dbe95e299a001ac7f361424036928750e97c017c4da6
3
- size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef04b50f717ccc1fb7d4248cbebe9f0c3884cadeb50b356efd9000eb189709aa
3
+ size 1383
checkpoints/checkpoint-94/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0cfd2d7ed5d052f6462cd13ac4e880e53bdfba79dce5db3e0ff53064d3a3eb5
3
- size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1053ab1e5d956431f3cae8a79e6dea26044c5acb288eb4cff4437e289c20048
3
+ size 1465
checkpoints/checkpoint-94/special_tokens_map.json CHANGED
@@ -1,10 +1,4 @@
1
  {
2
- "additional_special_tokens": [
3
- "▁<PRE>",
4
- "▁<MID>",
5
- "▁<SUF>",
6
- "▁<EOT>"
7
- ],
8
  "bos_token": {
9
  "content": "<s>",
10
  "lstrip": false,
@@ -19,7 +13,13 @@
19
  "rstrip": false,
20
  "single_word": false
21
  },
22
- "pad_token": "</s>",
 
 
 
 
 
 
23
  "unk_token": {
24
  "content": "<unk>",
25
  "lstrip": false,
 
1
  {
 
 
 
 
 
 
2
  "bos_token": {
3
  "content": "<s>",
4
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
  "unk_token": {
24
  "content": "<unk>",
25
  "lstrip": false,
checkpoints/checkpoint-94/tokenizer.json CHANGED
@@ -10,7 +10,7 @@
10
  "strategy": {
11
  "Fixed": 512
12
  },
13
- "direction": "Left",
14
  "pad_to_multiple_of": null,
15
  "pad_id": 2,
16
  "pad_type_id": 0,
@@ -43,42 +43,6 @@
43
  "rstrip": false,
44
  "normalized": false,
45
  "special": true
46
- },
47
- {
48
- "id": 32007,
49
- "content": "▁<PRE>",
50
- "single_word": false,
51
- "lstrip": false,
52
- "rstrip": false,
53
- "normalized": false,
54
- "special": true
55
- },
56
- {
57
- "id": 32008,
58
- "content": "▁<SUF>",
59
- "single_word": false,
60
- "lstrip": false,
61
- "rstrip": false,
62
- "normalized": false,
63
- "special": true
64
- },
65
- {
66
- "id": 32009,
67
- "content": "▁<MID>",
68
- "single_word": false,
69
- "lstrip": false,
70
- "rstrip": false,
71
- "normalized": false,
72
- "special": true
73
- },
74
- {
75
- "id": 32010,
76
- "content": "▁<EOT>",
77
- "single_word": false,
78
- "lstrip": false,
79
- "rstrip": false,
80
- "normalized": false,
81
- "special": true
82
  }
83
  ],
84
  "normalizer": {
@@ -32185,23 +32149,7 @@
32185
  "μ™•": 31996,
32186
  "ζ”Ά": 31997,
32187
  "弘": 31998,
32188
- "η»™": 31999,
32189
- "▁<SU": 32000,
32190
- "▁<SUF": 32001,
32191
- "▁<PRE": 32002,
32192
- "▁<M": 32003,
32193
- "▁<MID": 32004,
32194
- "▁<E": 32005,
32195
- "▁<EOT": 32006,
32196
- "▁<PRE>": 32007,
32197
- "▁<SUF>": 32008,
32198
- "▁<MID>": 32009,
32199
- "▁<EOT>": 32010,
32200
- "▁<EOT><EOT>": 32011,
32201
- "▁<EOT><EOT><EOT>": 32012,
32202
- "▁<EOT><EOT><EOT><EOT>": 32013,
32203
- "▁<EOT><EOT><EOT><EOT><EOT>": 32014,
32204
- "▁<EOT><EOT><EOT><EOT><EOT><EOT>": 32015
32205
  },
32206
  "merges": [
32207
  [
@@ -277199,50 +277147,6 @@
277199
  [
277200
  "▁",
277201
  "▁▁▁▁▁▁▁▁▁▁▁▁▁▁"
277202
- ],
277203
- [
277204
- "▁<",
277205
- "SU"
277206
- ],
277207
- [
277208
- "▁<SU",
277209
- "F"
277210
- ],
277211
- [
277212
- "▁<",
277213
- "PRE"
277214
- ],
277215
- [
277216
- "▁<",
277217
- "M"
277218
- ],
277219
- [
277220
- "▁<M",
277221
- "ID"
277222
- ],
277223
- [
277224
- "▁<",
277225
- "E"
277226
- ],
277227
- [
277228
- "▁<E",
277229
- "OT"
277230
- ],
277231
- [
277232
- "▁<PRE",
277233
- ">"
277234
- ],
277235
- [
277236
- "▁<SUF",
277237
- ">"
277238
- ],
277239
- [
277240
- "▁<MID",
277241
- ">"
277242
- ],
277243
- [
277244
- "▁<EOT",
277245
- ">"
277246
  ]
277247
  ]
277248
  }
 
10
  "strategy": {
11
  "Fixed": 512
12
  },
13
+ "direction": "Right",
14
  "pad_to_multiple_of": null,
15
  "pad_id": 2,
16
  "pad_type_id": 0,
 
43
  "rstrip": false,
44
  "normalized": false,
45
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
46
  }
47
  ],
48
  "normalizer": {
 
32149
  "μ™•": 31996,
32150
  "ζ”Ά": 31997,
32151
  "弘": 31998,
32152
+ "η»™": 31999
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32153
  },
32154
  "merges": [
32155
  [
 
277147
  [
277148
  "▁",
277149
  "▁▁▁▁▁▁▁▁▁▁▁▁▁▁"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
277150
  ]
277151
  ]
277152
  }
checkpoints/checkpoint-94/tokenizer.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:45ccb9c8b6b561889acea59191d66986d314e7cbd6a78abc6e49b139ca91c1e6
3
- size 500058
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
checkpoints/checkpoint-94/tokenizer_config.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "add_bos_token": true,
3
  "add_eos_token": false,
 
4
  "added_tokens_decoder": {
5
  "0": {
6
  "content": "<unk>",
@@ -25,60 +26,18 @@
25
  "rstrip": false,
26
  "single_word": false,
27
  "special": true
28
- },
29
- "32007": {
30
- "content": "▁<PRE>",
31
- "lstrip": false,
32
- "normalized": false,
33
- "rstrip": false,
34
- "single_word": false,
35
- "special": true
36
- },
37
- "32008": {
38
- "content": "▁<SUF>",
39
- "lstrip": false,
40
- "normalized": false,
41
- "rstrip": false,
42
- "single_word": false,
43
- "special": true
44
- },
45
- "32009": {
46
- "content": "▁<MID>",
47
- "lstrip": false,
48
- "normalized": false,
49
- "rstrip": false,
50
- "single_word": false,
51
- "special": true
52
- },
53
- "32010": {
54
- "content": "▁<EOT>",
55
- "lstrip": false,
56
- "normalized": false,
57
- "rstrip": false,
58
- "single_word": false,
59
- "special": true
60
  }
61
  },
62
- "additional_special_tokens": [
63
- "▁<PRE>",
64
- "▁<MID>",
65
- "▁<SUF>",
66
- "▁<EOT>"
67
- ],
68
  "bos_token": "<s>",
69
  "clean_up_tokenization_spaces": false,
70
  "eos_token": "</s>",
71
- "eot_token": "▁<EOT>",
72
  "extra_special_tokens": {},
73
- "fill_token": "<FILL_ME>",
74
- "legacy": null,
75
- "middle_token": "▁<MID>",
76
- "model_max_length": 1000000000000000019884624838656,
77
  "pad_token": "</s>",
78
- "prefix_token": "▁<PRE>",
79
  "sp_model_kwargs": {},
80
- "suffix_token": "▁<SUF>",
81
- "tokenizer_class": "CodeLlamaTokenizer",
82
  "unk_token": "<unk>",
83
  "use_default_system_prompt": false
84
  }
 
1
  {
2
  "add_bos_token": true,
3
  "add_eos_token": false,
4
+ "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
  "0": {
7
  "content": "<unk>",
 
26
  "rstrip": false,
27
  "single_word": false,
28
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  }
30
  },
 
 
 
 
 
 
31
  "bos_token": "<s>",
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
 
34
  "extra_special_tokens": {},
35
+ "legacy": false,
36
+ "model_max_length": 2048,
 
 
37
  "pad_token": "</s>",
38
+ "padding_side": "right",
39
  "sp_model_kwargs": {},
40
+ "tokenizer_class": "LlamaTokenizer",
 
41
  "unk_token": "<unk>",
42
  "use_default_system_prompt": false
43
  }
checkpoints/checkpoint-94/trainer_state.json CHANGED
@@ -11,30 +11,30 @@
11
  "log_history": [
12
  {
13
  "epoch": 0.42780748663101603,
14
- "grad_norm": 0.17147783935070038,
15
  "learning_rate": 0.00015957446808510637,
16
- "loss": 12.7183,
17
  "step": 20
18
  },
19
  {
20
  "epoch": 0.8556149732620321,
21
- "grad_norm": 0.10348663479089737,
22
  "learning_rate": 0.00011702127659574468,
23
- "loss": 12.9983,
24
  "step": 40
25
  },
26
  {
27
  "epoch": 1.2780748663101604,
28
- "grad_norm": 0.07714056968688965,
29
  "learning_rate": 7.446808510638298e-05,
30
- "loss": 12.7687,
31
  "step": 60
32
  },
33
  {
34
  "epoch": 1.7058823529411766,
35
- "grad_norm": 0.09477601200342178,
36
  "learning_rate": 3.191489361702128e-05,
37
- "loss": 12.7049,
38
  "step": 80
39
  }
40
  ],
@@ -55,7 +55,7 @@
55
  "attributes": {}
56
  }
57
  },
58
- "total_flos": 1.5192512670990336e+16,
59
  "train_batch_size": 2,
60
  "trial_name": null,
61
  "trial_params": null
 
11
  "log_history": [
12
  {
13
  "epoch": 0.42780748663101603,
14
+ "grad_norm": 5.215646743774414,
15
  "learning_rate": 0.00015957446808510637,
16
+ "loss": 6.4825,
17
  "step": 20
18
  },
19
  {
20
  "epoch": 0.8556149732620321,
21
+ "grad_norm": 0.32128605246543884,
22
  "learning_rate": 0.00011702127659574468,
23
+ "loss": 0.3244,
24
  "step": 40
25
  },
26
  {
27
  "epoch": 1.2780748663101604,
28
+ "grad_norm": 0.23809807002544403,
29
  "learning_rate": 7.446808510638298e-05,
30
+ "loss": 0.2467,
31
  "step": 60
32
  },
33
  {
34
  "epoch": 1.7058823529411766,
35
+ "grad_norm": 0.2185528576374054,
36
  "learning_rate": 3.191489361702128e-05,
37
+ "loss": 0.2283,
38
  "step": 80
39
  }
40
  ],
 
55
  "attributes": {}
56
  }
57
  },
58
+ "total_flos": 2379748793647104.0,
59
  "train_batch_size": 2,
60
  "trial_name": null,
61
  "trial_params": null
checkpoints/checkpoint-94/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d5668cf3ad716b632c2239a3e191b3db8db0d2224ecfa266312785c95ce7e77a
3
- size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a0a2a708a75961ab713089d6d62ba16b5b57b846e10dcd4ca4a7b443e0d86af
3
+ size 5777
final_adapter/README.md CHANGED
@@ -1,9 +1,9 @@
1
  ---
2
- base_model: codellama/CodeLlama-7b-hf
3
  library_name: peft
4
  pipeline_tag: text-generation
5
  tags:
6
- - base_model:adapter:codellama/CodeLlama-7b-hf
7
  - lora
8
  - transformers
9
  ---
 
1
  ---
2
+ base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
3
  library_name: peft
4
  pipeline_tag: text-generation
5
  tags:
6
+ - base_model:adapter:TinyLlama/TinyLlama-1.1B-Chat-v1.0
7
  - lora
8
  - transformers
9
  ---
final_adapter/adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "codellama/CodeLlama-7b-hf",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
@@ -25,8 +25,8 @@
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
28
- "v_proj",
29
- "q_proj"
30
  ],
31
  "target_parameters": null,
32
  "task_type": "CAUSAL_LM",
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
 
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
28
+ "q_proj",
29
+ "v_proj"
30
  ],
31
  "target_parameters": null,
32
  "task_type": "CAUSAL_LM",
final_adapter/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a6ad631c5d6c6d209ebd70a23c865e4b110c827017f205b18a2432ebbc554f3f
3
- size 16794200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8f402414502e7546231249c382c071c163a58eec09838b31328d00db17f99cf
3
+ size 4517152
final_adapter/chat_template.jinja ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {% for message in messages %}
2
+ {% if message['role'] == 'user' %}
3
+ {{ '<|user|>
4
+ ' + message['content'] + eos_token }}
5
+ {% elif message['role'] == 'system' %}
6
+ {{ '<|system|>
7
+ ' + message['content'] + eos_token }}
8
+ {% elif message['role'] == 'assistant' %}
9
+ {{ '<|assistant|>
10
+ ' + message['content'] + eos_token }}
11
+ {% endif %}
12
+ {% if loop.last and add_generation_prompt %}
13
+ {{ '<|assistant|>' }}
14
+ {% endif %}
15
+ {% endfor %}
final_adapter/special_tokens_map.json CHANGED
@@ -1,10 +1,4 @@
1
  {
2
- "additional_special_tokens": [
3
- "▁<PRE>",
4
- "▁<MID>",
5
- "▁<SUF>",
6
- "▁<EOT>"
7
- ],
8
  "bos_token": {
9
  "content": "<s>",
10
  "lstrip": false,
@@ -19,7 +13,13 @@
19
  "rstrip": false,
20
  "single_word": false
21
  },
22
- "pad_token": "</s>",
 
 
 
 
 
 
23
  "unk_token": {
24
  "content": "<unk>",
25
  "lstrip": false,
 
1
  {
 
 
 
 
 
 
2
  "bos_token": {
3
  "content": "<s>",
4
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
  "unk_token": {
24
  "content": "<unk>",
25
  "lstrip": false,
final_adapter/tokenizer.json CHANGED
@@ -10,7 +10,7 @@
10
  "strategy": {
11
  "Fixed": 512
12
  },
13
- "direction": "Left",
14
  "pad_to_multiple_of": null,
15
  "pad_id": 2,
16
  "pad_type_id": 0,
@@ -43,42 +43,6 @@
43
  "rstrip": false,
44
  "normalized": false,
45
  "special": true
46
- },
47
- {
48
- "id": 32007,
49
- "content": "▁<PRE>",
50
- "single_word": false,
51
- "lstrip": false,
52
- "rstrip": false,
53
- "normalized": false,
54
- "special": true
55
- },
56
- {
57
- "id": 32008,
58
- "content": "▁<SUF>",
59
- "single_word": false,
60
- "lstrip": false,
61
- "rstrip": false,
62
- "normalized": false,
63
- "special": true
64
- },
65
- {
66
- "id": 32009,
67
- "content": "▁<MID>",
68
- "single_word": false,
69
- "lstrip": false,
70
- "rstrip": false,
71
- "normalized": false,
72
- "special": true
73
- },
74
- {
75
- "id": 32010,
76
- "content": "▁<EOT>",
77
- "single_word": false,
78
- "lstrip": false,
79
- "rstrip": false,
80
- "normalized": false,
81
- "special": true
82
  }
83
  ],
84
  "normalizer": {
@@ -32185,23 +32149,7 @@
32185
  "μ™•": 31996,
32186
  "ζ”Ά": 31997,
32187
  "弘": 31998,
32188
- "η»™": 31999,
32189
- "▁<SU": 32000,
32190
- "▁<SUF": 32001,
32191
- "▁<PRE": 32002,
32192
- "▁<M": 32003,
32193
- "▁<MID": 32004,
32194
- "▁<E": 32005,
32195
- "▁<EOT": 32006,
32196
- "▁<PRE>": 32007,
32197
- "▁<SUF>": 32008,
32198
- "▁<MID>": 32009,
32199
- "▁<EOT>": 32010,
32200
- "▁<EOT><EOT>": 32011,
32201
- "▁<EOT><EOT><EOT>": 32012,
32202
- "▁<EOT><EOT><EOT><EOT>": 32013,
32203
- "▁<EOT><EOT><EOT><EOT><EOT>": 32014,
32204
- "▁<EOT><EOT><EOT><EOT><EOT><EOT>": 32015
32205
  },
32206
  "merges": [
32207
  [
@@ -277199,50 +277147,6 @@
277199
  [
277200
  "▁",
277201
  "▁▁▁▁▁▁▁▁▁▁▁▁▁▁"
277202
- ],
277203
- [
277204
- "▁<",
277205
- "SU"
277206
- ],
277207
- [
277208
- "▁<SU",
277209
- "F"
277210
- ],
277211
- [
277212
- "▁<",
277213
- "PRE"
277214
- ],
277215
- [
277216
- "▁<",
277217
- "M"
277218
- ],
277219
- [
277220
- "▁<M",
277221
- "ID"
277222
- ],
277223
- [
277224
- "▁<",
277225
- "E"
277226
- ],
277227
- [
277228
- "▁<E",
277229
- "OT"
277230
- ],
277231
- [
277232
- "▁<PRE",
277233
- ">"
277234
- ],
277235
- [
277236
- "▁<SUF",
277237
- ">"
277238
- ],
277239
- [
277240
- "▁<MID",
277241
- ">"
277242
- ],
277243
- [
277244
- "▁<EOT",
277245
- ">"
277246
  ]
277247
  ]
277248
  }
 
10
  "strategy": {
11
  "Fixed": 512
12
  },
13
+ "direction": "Right",
14
  "pad_to_multiple_of": null,
15
  "pad_id": 2,
16
  "pad_type_id": 0,
 
43
  "rstrip": false,
44
  "normalized": false,
45
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
46
  }
47
  ],
48
  "normalizer": {
 
32149
  "μ™•": 31996,
32150
  "ζ”Ά": 31997,
32151
  "弘": 31998,
32152
+ "η»™": 31999
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32153
  },
32154
  "merges": [
32155
  [
 
277147
  [
277148
  "▁",
277149
  "▁▁▁▁▁▁▁▁▁▁▁▁▁▁"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
277150
  ]
277151
  ]
277152
  }
final_adapter/tokenizer.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:45ccb9c8b6b561889acea59191d66986d314e7cbd6a78abc6e49b139ca91c1e6
3
- size 500058
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
final_adapter/tokenizer_config.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "add_bos_token": true,
3
  "add_eos_token": false,
 
4
  "added_tokens_decoder": {
5
  "0": {
6
  "content": "<unk>",
@@ -25,60 +26,18 @@
25
  "rstrip": false,
26
  "single_word": false,
27
  "special": true
28
- },
29
- "32007": {
30
- "content": "▁<PRE>",
31
- "lstrip": false,
32
- "normalized": false,
33
- "rstrip": false,
34
- "single_word": false,
35
- "special": true
36
- },
37
- "32008": {
38
- "content": "▁<SUF>",
39
- "lstrip": false,
40
- "normalized": false,
41
- "rstrip": false,
42
- "single_word": false,
43
- "special": true
44
- },
45
- "32009": {
46
- "content": "▁<MID>",
47
- "lstrip": false,
48
- "normalized": false,
49
- "rstrip": false,
50
- "single_word": false,
51
- "special": true
52
- },
53
- "32010": {
54
- "content": "▁<EOT>",
55
- "lstrip": false,
56
- "normalized": false,
57
- "rstrip": false,
58
- "single_word": false,
59
- "special": true
60
  }
61
  },
62
- "additional_special_tokens": [
63
- "▁<PRE>",
64
- "▁<MID>",
65
- "▁<SUF>",
66
- "▁<EOT>"
67
- ],
68
  "bos_token": "<s>",
69
  "clean_up_tokenization_spaces": false,
70
  "eos_token": "</s>",
71
- "eot_token": "▁<EOT>",
72
  "extra_special_tokens": {},
73
- "fill_token": "<FILL_ME>",
74
- "legacy": null,
75
- "middle_token": "▁<MID>",
76
- "model_max_length": 1000000000000000019884624838656,
77
  "pad_token": "</s>",
78
- "prefix_token": "▁<PRE>",
79
  "sp_model_kwargs": {},
80
- "suffix_token": "▁<SUF>",
81
- "tokenizer_class": "CodeLlamaTokenizer",
82
  "unk_token": "<unk>",
83
  "use_default_system_prompt": false
84
  }
 
1
  {
2
  "add_bos_token": true,
3
  "add_eos_token": false,
4
+ "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
  "0": {
7
  "content": "<unk>",
 
26
  "rstrip": false,
27
  "single_word": false,
28
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  }
30
  },
 
 
 
 
 
 
31
  "bos_token": "<s>",
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
 
34
  "extra_special_tokens": {},
35
+ "legacy": false,
36
+ "model_max_length": 2048,
 
 
37
  "pad_token": "</s>",
38
+ "padding_side": "right",
39
  "sp_model_kwargs": {},
40
+ "tokenizer_class": "LlamaTokenizer",
 
41
  "unk_token": "<unk>",
42
  "use_default_system_prompt": false
43
  }