sayakpaul HF Staff commited on
Commit
bff8dbb
·
verified ·
1 Parent(s): edac16b

manually update the torch dtype and see if that works!

Browse files
Files changed (1) hide show
  1. config.json +1 -1
config.json CHANGED
@@ -41,7 +41,7 @@
41
  "use_cache": true,
42
  "vocab_size": 262208
43
  },
44
- "torch_dtype": "bfloat16",
45
  "transformers_version": "4.50.0.dev0",
46
  "vision_config": {
47
  "attention_dropout": 0.0,
 
41
  "use_cache": true,
42
  "vocab_size": 262208
43
  },
44
+ "torch_dtype": "float32",
45
  "transformers_version": "4.50.0.dev0",
46
  "vision_config": {
47
  "attention_dropout": 0.0,