snx999 commited on
Commit
b1fd232
·
verified ·
1 Parent(s): ff682a5

Add files using upload-large-folder tool

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "./model-weights-10",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -22,6 +22,7 @@
22
  "num_attention_heads": 32,
23
  "num_hidden_layers": 32,
24
  "num_key_value_heads": 8,
 
25
  "pretraining_tp": 1,
26
  "rms_norm_eps": 1e-05,
27
  "rope_scaling": {
@@ -34,7 +35,7 @@
34
  "rope_theta": 500000.0,
35
  "tie_word_embeddings": false,
36
  "torch_dtype": "bfloat16",
37
- "transformers_version": "4.47.1",
38
- "use_cache": true,
39
  "vocab_size": 128257
40
  }
 
1
  {
2
+ "_name_or_path": "meta-llama/Llama-3.1-8B-Instruct",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
22
  "num_attention_heads": 32,
23
  "num_hidden_layers": 32,
24
  "num_key_value_heads": 8,
25
+ "pad_token_id": 128256,
26
  "pretraining_tp": 1,
27
  "rms_norm_eps": 1e-05,
28
  "rope_scaling": {
 
35
  "rope_theta": 500000.0,
36
  "tie_word_embeddings": false,
37
  "torch_dtype": "bfloat16",
38
+ "transformers_version": "4.48.0",
39
+ "use_cache": false,
40
  "vocab_size": 128257
41
  }
generation_config.json CHANGED
@@ -8,5 +8,5 @@
8
  ],
9
  "temperature": 0.6,
10
  "top_p": 0.9,
11
- "transformers_version": "4.47.1"
12
  }
 
8
  ],
9
  "temperature": 0.6,
10
  "top_p": 0.9,
11
+ "transformers_version": "4.48.0"
12
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a316cb8ec49e8d42de67ffaa5dadb9cbc1a8d380f16991eb0ca4a57a67cbafba
3
  size 4976706864
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7784df4a8c13f5813c4dfcc993157d8495f431564af417d92742de4a76890d5b
3
  size 4976706864
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:94248bd438d83ca7f3033f4080373f44a7839e749443134441fea84b3281db48
3
  size 4999802720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6cb748ba3699a707eb27acc8ab85a1ae91df16b440308bc4ab8e6b98fd637fb
3
  size 4999802720
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e33989c3547fa2575429ca0b2578c518c67433808f2a4cc9ab1bc659450da425
3
  size 4915916176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af706cd206454f677b2d453759a00b86f2f6ab5f862b1e9579877f694d0eabac
3
  size 4915916176
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2fef7b40c49a3d3521d4d29a111323ee6954a057183bf05a17618dc110c528ec
3
  size 1168147000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ea47ac39f025bf3373986be7d7e0195836ae266f079bde44bc2941b24308721
3
  size 1168147000
special_tokens_map.json CHANGED
@@ -14,7 +14,7 @@
14
  "single_word": false
15
  },
16
  "pad_token": {
17
- "content": "[PAD]",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
 
14
  "single_word": false
15
  },
16
  "pad_token": {
17
+ "content": "<|pad|>",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9d7cacaa40afe2956f08737f84e63925c473cf6675d90dfc8caeae75768f9b7
3
- size 17210372
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc5fe1965e6af7516a0d52cadb47bc6bd7789e6cf6d305a300d2129281e5cbd0
3
+ size 17210376
tokenizer_config.json CHANGED
@@ -2049,7 +2049,7 @@
2049
  "special": true
2050
  },
2051
  "128256": {
2052
- "content": "[PAD]",
2053
  "lstrip": false,
2054
  "normalized": false,
2055
  "rstrip": false,
@@ -2067,6 +2067,6 @@
2067
  "attention_mask"
2068
  ],
2069
  "model_max_length": 131072,
2070
- "pad_token": "[PAD]",
2071
  "tokenizer_class": "PreTrainedTokenizerFast"
2072
  }
 
2049
  "special": true
2050
  },
2051
  "128256": {
2052
+ "content": "<|pad|>",
2053
  "lstrip": false,
2054
  "normalized": false,
2055
  "rstrip": false,
 
2067
  "attention_mask"
2068
  ],
2069
  "model_max_length": 131072,
2070
+ "pad_token": "<|pad|>",
2071
  "tokenizer_class": "PreTrainedTokenizerFast"
2072
  }