Update README.md
Browse files
README.md
CHANGED
|
@@ -49,7 +49,7 @@ import torch
|
|
| 49 |
# Load model with FP8 quantization
|
| 50 |
model = AutoModelForCausalLM.from_pretrained(
|
| 51 |
"TevunahAi/NextCoder-7B-FP8",
|
| 52 |
-
torch_dtype=torch.
|
| 53 |
device_map="auto",
|
| 54 |
low_cpu_mem_usage=True,
|
| 55 |
)
|
|
|
|
| 49 |
# Load model with FP8 quantization
|
| 50 |
model = AutoModelForCausalLM.from_pretrained(
|
| 51 |
"TevunahAi/NextCoder-7B-FP8",
|
| 52 |
+
torch_dtype=torch.bfloat16,
|
| 53 |
device_map="auto",
|
| 54 |
low_cpu_mem_usage=True,
|
| 55 |
)
|