[Paper Reading] PENCIL: Long Thoughts with Short Memory
π
For mobile: View PDF
Main References
- Chenxiao Yang, Nathan Srebro, David McAllester, and Zhiyuan Li. PENCIL: Long Thoughts with Short Memory. arXiv preprint, 2025.
Supplementary References
- Maxwell Nye et al. Show Your Work: Scratchpads for Intermediate Computation with Language Models. arXiv preprint, 2021.
- Jason Wei et al. Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. NeurIPS 2022.
- William Merrill and Ashish Sabharwal, The Parallelism Tradeoff: Limitations of Log-Precision Transformers. TACL 2023.
- William Merrill and Ashish Sabharwal. The Expressive Power of Transformers with Chain of Thought. ICLR 2024.