zainabkhan commited on
Commit
67c5a4c
·
verified ·
1 Parent(s): 25d05e0

Upload PhiForCausalLM

Browse files
Files changed (2) hide show
  1. config.json +1 -1
  2. model.safetensors +1 -1
config.json CHANGED
@@ -27,7 +27,7 @@
27
  "quantization_config": {
28
  "_load_in_4bit": true,
29
  "_load_in_8bit": false,
30
- "bnb_4bit_compute_dtype": "float16",
31
  "bnb_4bit_quant_storage": "uint8",
32
  "bnb_4bit_quant_type": "nf4",
33
  "bnb_4bit_use_double_quant": false,
 
27
  "quantization_config": {
28
  "_load_in_4bit": true,
29
  "_load_in_8bit": false,
30
+ "bnb_4bit_compute_dtype": "bfloat16",
31
  "bnb_4bit_quant_storage": "uint8",
32
  "bnb_4bit_quant_type": "nf4",
33
  "bnb_4bit_use_double_quant": false,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9183558b8f5d144803ec41b9dfd0cdbea0d81e7c870826fe287833bd275b1245
3
  size 2132377992
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ee7ab92c1b48735ca7e933cdce08f6514be4cc61c9621e8791d7aa618f8e0d4
3
  size 2132377992