ReasoningQAT
Collection
6 items • Updated
This model is a 2-bit pseudo-quantized version of Qwen3-8B, trained with Quantization-Aware Training (QAT) for reasoning tasks.
@inproceedings{
okoshi2026towards,
title={Towards Quantization-Aware Training for Ultra-Low-Bit Reasoning {LLM}s},
author={Yasuyuki Okoshi and Hikari Otsuka and Daichi Fujiki and Masato Motomura},
booktitle={The Fourteenth International Conference on Learning Representations},
year={2026},
url={https://openreview.net/forum?id=Azsd2qyK6C}
}