filipemiguelmartins commited on
Commit
e2ea3a9
·
verified ·
1 Parent(s): 03d793f

Upload LlamaForCausalLM

Browse files
config.json CHANGED
@@ -25,5 +25,5 @@
25
  "torch_dtype": "bfloat16",
26
  "transformers_version": "4.41.2",
27
  "use_cache": true,
28
- "vocab_size": 128264
29
  }
 
25
  "torch_dtype": "bfloat16",
26
  "transformers_version": "4.41.2",
27
  "use_cache": true,
28
+ "vocab_size": 128256
29
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:11a795fedb82ce808b73a52b73a0e1acc02b04178f0c763a6b3c8cfaa4886010
3
- size 4976764208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8f0a69fe78730bb9f8e75ebe279d450e268c43276874e27489843e47398bd06
3
+ size 4976698672
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c296b1024f077c3a4b558a3d57cc7c3bf8a8d7148999eeefe11eaf8f58ce6afc
3
  size 4999802720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d49f6aad8b177e06ffa21c448288e66aada69afcefd85cde1fa3c3f98b2ed4c2
3
  size 4999802720
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:86adbb59d605076458a1f05ed3c23298c6585da8bb10f4e6aaaa9820def88258
3
  size 4915916176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e58079b5e019b747713ab713fec01124b85bf5e4b351890e94f88d305057fe7
3
  size 4915916176
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2dc296a5aafdecf3a1e1414c2590a247fe8aa733d01cafe6995cd818b0ca05ec
3
- size 1168204344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1844a5727bcea3f8f12c7ce9cfd799ce4a583a9dcc949d91a6295f52d0647563
3
+ size 1168138808
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 16060653568
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00004-of-00004.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 16060522496
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00004-of-00004.safetensors",