qwen3-8b-dmax

DMax/OPUT-trained variant of Qwen/Qwen3-8B.

  • Base: Qwen/Qwen3-8B
  • Training framework: dllm-jax (JAX/Flax NNX, TPU)
  • Source checkpoint: gs://dllm-jax-europe-west4/checkpoints/qwen3-8b-PRODUCTION-mix-dmax-opUT-optionB-v6e64-20260504-0500/checkpoint_1000
  • Step: 1000

This is a JAX-trained DMax/OPUT (block-diffusion / on-policy under-tuning) fine-tune. The weights are stored in standard HF Qwen3ForCausalLM layout โ€” the same loader path as the base model. Inference requires the dllm-jax DMax block-diffusion path (the model is not an autoregressive Qwen3 โ€” it expects a doubled [noised; clean] input under a block-diffusion mask).

Downloads last month
684
Safetensors
Model size
8B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for beomi/qwen3-8b-dmax

Finetuned
Qwen/Qwen3-8B
Finetuned
(1568)
this model