Text Generation
Transformers
Safetensors
mistral
nvfp4
modelopt
quantized
blackwell
b200
conversational
text-generation-inference
8-bit precision
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("TheHouseOfTheDude/Behemoth-R1-V2_ModelOpt-NVFP4")
model = AutoModelForCausalLM.from_pretrained("TheHouseOfTheDude/Behemoth-R1-V2_ModelOpt-NVFP4")
messages = [
{"role": "user", "content": "Who are you?"},
]
inputs = tokenizer.apply_chat_template(
messages,
add_generation_prompt=True,
tokenize=True,
return_dict=True,
return_tensors="pt",
).to(model.device)
outputs = model.generate(**inputs, max_new_tokens=40)
print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:]))Quick Links
Behemoth-R1-V2 ModelOpt NVFP4
NVFP4 quantized version of TheDrummer/Behemoth-R1-123B-v2 using NVIDIA Model Optimizer.
Quantization Details
| Property | Value |
|---|---|
| Original Model | TheDrummer/Behemoth-R1-123B-v2 |
| Quantization | NVFP4 (FP4 weights, FP16 activations) |
| Method | NVIDIA ModelOpt PTQ |
| Calibration Samples | 512 |
| Max Sequence Length | 4096 |
Hardware Requirements
- Optimal: NVIDIA Blackwell GPUs (B100, B200, RTX PRO 6000 Blackwell)
- Compatible: Hopper/Ampere (will use weight-only mode)
Usage with vLLM
from vllm import LLM, SamplingParams
llm = LLM(
model="TheHouseOfTheDude/Behemoth-R1-V2_ModelOpt-NVFP4",
quantization="modelopt",
trust_remote_code=True,
)
sampling_params = SamplingParams(temperature=0.8, top_p=0.95, max_tokens=512)
outputs = llm.generate(["Write a story about..."], sampling_params)
print(outputs[0].outputs[0].text)
Chat Template
Uses Mistral v7 (Non-Tekken) format. See the original model card for usage details.
Credits
- Original Model: TheDrummer
- Quantization: TheHouseOfTheDude
- Quantization Framework: NVIDIA ModelOpt
- Downloads last month
- 31
Model tree for TheHouseOfTheDude/Behemoth-R1-V2_ModelOpt-NVFP4
Base model
mistralai/Mistral-Large-Instruct-2411 Finetuned
TheDrummer/Behemoth-R1-123B-v2
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="TheHouseOfTheDude/Behemoth-R1-V2_ModelOpt-NVFP4") messages = [ {"role": "user", "content": "Who are you?"}, ] pipe(messages)