Light-IF-4B-f32-GGUF

Light-IF-4B is a 4-billion-parameter instruction-following language model derived from Qwen3-4B-Base, designed to overcome "lazy reasoning" in complex tasks by incorporating previewing and self-checking during inference; it is fine-tuned using entropy-preserving supervised learning (Entropy-SFT) and token-wise entropy-adaptive reinforcement learning (TEA-RL) on a carefully filtered dataset, producing strong results across instruction-following and reasoning benchmarks (such as SuperClue, IFEval, and IFBench), where it matches or outperforms even larger or closed-source models, and supports advanced features such as extended context (32k-131k tokens with YaRN), efficient deployment (via Hugging Face Transformers, sglang, or vllm), and open integration for research in robust generalizable reasoning, with further details, evaluation code, and licensing on its official Hugging Face repository and paper.

Model Files

File Name Size Quant Type
Qwen3-4B-MegaScience.BF16.gguf 8.05 GB BF16
Qwen3-4B-MegaScience.F16.gguf 8.05 GB F16
Qwen3-4B-MegaScience.F32.gguf 16.1 GB F32
Qwen3-4B-MegaScience.Q3_K_L.gguf 2.24 GB Q3_K_L
Qwen3-4B-MegaScience.Q3_K_S.gguf 1.89 GB Q3_K_S
Qwen3-4B-MegaScience.Q4_K_M.gguf 2.5 GB Q4_K_M
Qwen3-4B-MegaScience.Q4_K_S.gguf 2.38 GB Q4_K_S
Qwen3-4B-MegaScience.Q5_K_M.gguf 2.89 GB Q5_K_M
Qwen3-4B-MegaScience.Q5_K_S.gguf 2.82 GB Q5_K_S
Qwen3-4B-MegaScience.Q6_K.gguf 3.31 GB Q6_K
Qwen3-4B-MegaScience.Q8_0.gguf 4.28 GB Q8_0

Quants Usage

(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)

Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

image.png

Downloads last month
76
GGUF
Model size
4B params
Architecture
qwen3
Hardware compatibility
Log In to add your hardware

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

32-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for prithivMLmods/Light-IF-4B-f32-GGUF

Finetuned
Qwen/Qwen3-4B
Quantized
(3)
this model