File size: 142 Bytes
1fc92ce | 1 2 3 4 5 6 7 8 | {
"architectures": [
"AutoModelForCausalLM"
],
"model_type": "phi",
"transformers_version": "4.45.1",
"torch_dtype": "float16"
} |
1fc92ce | 1 2 3 4 5 6 7 8 | {
"architectures": [
"AutoModelForCausalLM"
],
"model_type": "phi",
"transformers_version": "4.45.1",
"torch_dtype": "float16"
} |