Rethinking External Slow-Thinking: From Snowball Errors to Probability of Correct Reasoning

21
citations
#257
in ICML 2025
of 3340 papers
3
Top Authors
4
Data Points

Abstract

Test-time scaling, which is also often referred to asslow-thinking, has been demonstrated to enhance multi-step reasoning in large language models (LLMs). However, despite its widespread utilization, the mechanisms underlying slow-thinking methods remain poorly understood. This paper explores the mechanisms of external slow-thinking from a theoretical standpoint. We begin by examining the snowball error effect within the LLM reasoning process and connect it to the likelihood of correct reasoning using information theory. Building on this, we show that external slow-thinking methods can be interpreted as strategies to mitigate the error probability. We further provide a comparative analysis of popular external slow-thinking approaches, ranging from simple to complex, highlighting their differences and interrelationships. Our findings suggest that the efficacy of these methods is not primarily determined by the specific framework employed, and that expanding the search scope or the model's internal reasoning capacity may yield more sustained improvements in the long term. We open-source our code at https://github.com/ZyGan1999/Snowball-Errors-and-Probability.

Citation History

Jan 28, 2026
0
Feb 13, 2026
21+21
Feb 13, 2026
21
Feb 13, 2026
21