SemCoT: Accelerating Chain-of-Thought Reasoning through Semantically-Aligned Implicit Tokens
Paper
β’
2510.24940
β’
Published
β’
18
SemCoT is a framework that improves the efficiency of Chain-of-Thought (CoT) reasoning by encoding reasoning steps inside hidden representations (implicit tokens) instead of generating long textual explanations. This approach significantly speeds up inference while maintaining high performance.
Specifically, this model uses princeton-nlp/Sheared-LLaMA-1.3B as the base model and is fine-tuned using the SemCoT framework on the tau/commonsense_qa dataset.
Please refer to the official GitHub repository for detailed instructions on how to load and use the model, as it utilizes custom hidden tokens for implicit reasoning.
@inproceedings{he2025semcot,
title={SemCoT: Accelerating Chain-of-Thought Reasoning through Semantically-Aligned Implicit Tokens},
author={He, Yinhan and Zheng, Wendy and Zhu, Yaochen and Zheng, Zaiyi and Su, Lin and Vasudevan, Sriram and Guo, Qi and Hong, Liangjie and Li, Jundong},
booktitle={39th Conference on Neural Information Processing Systems (NeurIPS 2025)},
year={2025}
}
Base model
princeton-nlp/Sheared-LLaMA-1.3B