Kfjjdjdjdhdhd commited on
Commit
c1f2402
·
verified ·
1 Parent(s): f0bbbdd

Upload PhiForCausalLM

Browse files
Files changed (2) hide show
  1. config.json +1 -1
  2. model.safetensors +1 -1
config.json CHANGED
@@ -14,7 +14,7 @@
14
  "intermediate_size": 8192,
15
  "layer_norm_eps": 1e-05,
16
  "layer_norm_epsilon": 1e-05,
17
- "model_type": "phi",
18
  "n_embd": 1024,
19
  "n_head": 16,
20
  "n_head_kv": null,
 
14
  "intermediate_size": 8192,
15
  "layer_norm_eps": 1e-05,
16
  "layer_norm_epsilon": 1e-05,
17
+ "model_type": "phi-causal-lm",
18
  "n_embd": 1024,
19
  "n_head": 16,
20
  "n_head_kv": null,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7879646e2b1e0e2b1ccfe9d67dfdfb3ed82c53fb452f54b7a7bcc2b301c09258
3
  size 469800176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59007ea68bc6da0d38849f27d8017082f443ba2d233b1fd521eeced240fce7fc
3
  size 469800176