SemCoT: Accelerating Chain-of-Thought Reasoning through Semantically-Aligned Implicit Tokens
Paper
β’ 2510.24940 β’ Published
β’ 18
SemCoT is a framework that improves the efficiency of Chain-of-Thought (CoT) reasoning by encoding reasoning steps inside hidden representations instead of generating long textual explanations. This implicit reasoning greatly speeds up inference while keeping performance high.
Specifically, this checkpoint is a fine-tuned version of princeton-nlp/Sheared-LLaMA-1.3B using the SemCoT framework on the skrishna/coin_flip dataset.
If you find this work useful, please cite our paper:
@inproceedings{he2025semcot,
title={SemCoT: Accelerating Chain-of-Thought Reasoning through Semantically-Aligned Implicit Tokens},
author={He, Yinhan and Zheng, Wendy and Zhu, Yaochen and Zheng, Zaiyi and Su, Lin and Vasudevan, Sriram and Guo, Qi and Hong, Liangjie and Li, Jundong},
booktitle={39th Conference on Neural Information Processing Systems (NeurIPS 2025)},
year={2025}
}
Base model
princeton-nlp/Sheared-LLaMA-1.3B