cipher982 commited on
Commit
2c96a7e
·
1 Parent(s): e018091

Upload LlamaForCausalLM

Browse files
config.json CHANGED
@@ -22,5 +22,5 @@
22
  "torch_dtype": "bfloat16",
23
  "transformers_version": "4.31.0",
24
  "use_cache": true,
25
- "vocab_size": 32004
26
  }
 
22
  "torch_dtype": "bfloat16",
23
  "transformers_version": "4.31.0",
24
  "use_cache": true,
25
+ "vocab_size": 32007
26
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a84844f5ed96975c5d0c41bd75604635248bb304d3226f16b45dfffab3163d65
3
- size 9976611912
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5af85aaaeb49b9245e446396ae77bae5ab02e483a78560670e968a572bb1a1fc
3
+ size 9976636488
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a8c18d5cb2c3b52426f9b4a459758d6db4d2d493a3387066881e042a9cc0324
3
- size 3500330192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fa9cf6f92e9a6c210ccfc229ac1935567f31be0581689d96cec6ce1ab371f93
3
+ size 3500354768
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 13476904960
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00002-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 13476954112
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00002-of-00002.safetensors",