|
|
--- |
|
|
language: |
|
|
- en |
|
|
license: apache-2.0 |
|
|
pipeline_tag: text-generation |
|
|
tags: |
|
|
- reasoning |
|
|
- looped transformer |
|
|
arxiv: 2511.08577 |
|
|
library_name: transformers |
|
|
datasets: |
|
|
- open-r1/Mixture-of-Thoughts |
|
|
base_model: |
|
|
- Qwen/Qwen3-1.7B-Base |
|
|
--- |
|
|
|
|
|
This is the general version of Standard-1.7B, trained on a mixture of math, code, and science data, presented in the paper [Think-at-Hard: Selective Latent Iterations to Improve Reasoning Language Models](https://huggingface.co/papers/2511.08577). |
|
|
|
|
|
Please visit our [GitHub repo](https://github.com/thu-nics/TaH) for more information. |
|
|
|
|
|
|
|
|
### Sample Usage |
|
|
|
|
|
Please see [Github Example](https://github.com/thu-nics/TaH?tab=readme-ov-file#run-an-example-for-tah) for sample usage. |