Triune commited on
Commit
b71ad53
·
1 Parent(s): 425732f

Upload 16 files

Browse files
_gitattributes ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/home/keyan/gpt/llm_model/WizardLM-13B-v1.0-hf",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -13,14 +13,11 @@
13
  "model_type": "llama",
14
  "num_attention_heads": 40,
15
  "num_hidden_layers": 40,
16
- "num_key_value_heads": 40,
17
  "pad_token_id": 0,
18
- "pretraining_tp": 1,
19
  "rms_norm_eps": 1e-06,
20
- "rope_scaling": null,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "float32",
23
- "transformers_version": "4.31.0",
24
  "use_cache": false,
25
  "vocab_size": 32001
26
  }
 
1
  {
2
+ "_name_or_path": "../../WizardLM-13B-V1.0",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
13
  "model_type": "llama",
14
  "num_attention_heads": 40,
15
  "num_hidden_layers": 40,
 
16
  "pad_token_id": 0,
 
17
  "rms_norm_eps": 1e-06,
 
18
  "tie_word_embeddings": false,
19
  "torch_dtype": "float32",
20
+ "transformers_version": "4.29.2",
21
  "use_cache": false,
22
  "vocab_size": 32001
23
  }
generation_config.json CHANGED
@@ -3,5 +3,5 @@
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
  "pad_token_id": 0,
6
- "transformers_version": "4.31.0"
7
  }
 
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
  "pad_token_id": 0,
6
+ "transformers_version": "4.29.2"
7
  }
pytorch_model-00001-of-00006.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:733b4bb53be931f8883118557aea5915e2833df25688946858a5bda94a435149
3
+ size 9956564363
pytorch_model-00002-of-00006.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66fe1d8f9d6ea1e74f499a4eb9b1d696987c1aa346205e7d8845217a0fafbd1e
3
+ size 9940856385
pytorch_model-00003-of-00006.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d63e2ed09e6768995483c2398eec832b3d35e2003158c9901f2ff2da910d1ff
3
+ size 9940856943
pytorch_model-00004-of-00006.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71dac6c7301d83b20a4ca1e1f8a8e07074ebf79d1121721804da642f8616a2da
3
+ size 9867415289
pytorch_model-00005-of-00006.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07981bf71e77bb2c62dc6b6bd4cd0fb65f8d4f7484adeb3bd1c75885a6ca3072
3
+ size 9867456961
pytorch_model-00006-of-00006.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9e7618baebd080fc7188c5b47a7d9286a04a60cdb33ee4626523f2bf998b8c7
3
+ size 2490496687
pytorch_model.bin.index.json CHANGED
@@ -246,9 +246,9 @@
246
  "model.layers.3.self_attn.rotary_emb.inv_freq": "pytorch_model-00001-of-00006.bin",
247
  "model.layers.3.self_attn.v_proj.weight": "pytorch_model-00001-of-00006.bin",
248
  "model.layers.30.input_layernorm.weight": "pytorch_model-00005-of-00006.bin",
249
- "model.layers.30.mlp.down_proj.weight": "pytorch_model-00005-of-00006.bin",
250
  "model.layers.30.mlp.gate_proj.weight": "pytorch_model-00004-of-00006.bin",
251
- "model.layers.30.mlp.up_proj.weight": "pytorch_model-00004-of-00006.bin",
252
  "model.layers.30.post_attention_layernorm.weight": "pytorch_model-00005-of-00006.bin",
253
  "model.layers.30.self_attn.k_proj.weight": "pytorch_model-00004-of-00006.bin",
254
  "model.layers.30.self_attn.o_proj.weight": "pytorch_model-00004-of-00006.bin",
 
246
  "model.layers.3.self_attn.rotary_emb.inv_freq": "pytorch_model-00001-of-00006.bin",
247
  "model.layers.3.self_attn.v_proj.weight": "pytorch_model-00001-of-00006.bin",
248
  "model.layers.30.input_layernorm.weight": "pytorch_model-00005-of-00006.bin",
249
+ "model.layers.30.mlp.down_proj.weight": "pytorch_model-00004-of-00006.bin",
250
  "model.layers.30.mlp.gate_proj.weight": "pytorch_model-00004-of-00006.bin",
251
+ "model.layers.30.mlp.up_proj.weight": "pytorch_model-00005-of-00006.bin",
252
  "model.layers.30.post_attention_layernorm.weight": "pytorch_model-00005-of-00006.bin",
253
  "model.layers.30.self_attn.k_proj.weight": "pytorch_model-00004-of-00006.bin",
254
  "model.layers.30.self_attn.o_proj.weight": "pytorch_model-00004-of-00006.bin",
tokenizer_config.json CHANGED
@@ -1,4 +1,6 @@
1
  {
 
 
2
  "bos_token": {
3
  "__type": "AddedToken",
4
  "content": "<s>",
 
1
  {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
  "bos_token": {
5
  "__type": "AddedToken",
6
  "content": "<s>",