researchvia ArXiv cs.AI

Reasoning Fails in Large Models

Large reasoning models perform well on multi-step tasks but have unstable behavior. Step-Saliency analysis reveals information-flow failures.

Reasoning Fails in Large Models

Researchers have introduced Step-Saliency, a tool to analyze large reasoning models (LRMs) that generate long chains of thought. This tool pools attention--gradient scores into step-to-step maps, providing insight into the models' behavior.

The analysis reveals two recurring information-flow failures: Shallow Lock-in, where shallow layers over-focus, and another failure related to the models' inability to effectively process and utilize information across different steps.

The introduction of Step-Saliency and its findings have implications for the development of more robust and interpretable LRMs. As the field continues to advance, tools like Step-Saliency will be crucial in identifying and addressing the limitations of these models, ultimately leading to more reliable and efficient reasoning systems.

#reasoning#large-models#analysis#interpretability#information-flow