EddySL commited on
Commit
394c684
·
verified ·
1 Parent(s): 42ab58a

(Trained with Unsloth)

Browse files
Files changed (1) hide show
  1. config.json +4 -4
config.json CHANGED
@@ -42,7 +42,7 @@
42
  2
43
  ]
44
  ],
45
- "torch_dtype": "bfloat16",
46
  "vocab_offset": 262272,
47
  "vocab_size": 128
48
  },
@@ -194,12 +194,12 @@
194
  "rope_scaling": null,
195
  "rope_theta": 1000000.0,
196
  "sliding_window": 512,
197
- "torch_dtype": "bfloat16",
198
  "use_cache": true,
199
  "vocab_size": 262400,
200
  "vocab_size_per_layer_input": 262144
201
  },
202
- "torch_dtype": "bfloat16",
203
  "transformers_version": "4.55.0",
204
  "unsloth_fixed": true,
205
  "unsloth_version": "2025.8.4",
@@ -216,7 +216,7 @@
216
  "model_type": "gemma3n_vision",
217
  "num_classes": 2,
218
  "rms_norm_eps": 1e-06,
219
- "torch_dtype": "bfloat16",
220
  "vocab_offset": 262144,
221
  "vocab_size": 128
222
  },
 
42
  2
43
  ]
44
  ],
45
+ "torch_dtype": "float16",
46
  "vocab_offset": 262272,
47
  "vocab_size": 128
48
  },
 
194
  "rope_scaling": null,
195
  "rope_theta": 1000000.0,
196
  "sliding_window": 512,
197
+ "torch_dtype": "float16",
198
  "use_cache": true,
199
  "vocab_size": 262400,
200
  "vocab_size_per_layer_input": 262144
201
  },
202
+ "torch_dtype": "float16",
203
  "transformers_version": "4.55.0",
204
  "unsloth_fixed": true,
205
  "unsloth_version": "2025.8.4",
 
216
  "model_type": "gemma3n_vision",
217
  "num_classes": 2,
218
  "rms_norm_eps": 1e-06,
219
+ "torch_dtype": "float16",
220
  "vocab_offset": 262144,
221
  "vocab_size": 128
222
  },