Papers
arxiv:2601.12538

Agentic Reasoning for Large Language Models

Published on Jan 18
ยท Submitted by
Jiaru Zou
on Jan 22
#1 Paper of the day
Authors:
,
,
,
,
,
,
,
,
,
,
,
,
,
,
,
,
,

Abstract

Agentic reasoning redefines large language models as autonomous agents capable of planning, acting, and learning through continuous interaction in dynamic environments across single-agent and multi-agent frameworks.

AI-generated summary

Reasoning is a fundamental cognitive process underlying inference, problem-solving, and decision-making. While large language models (LLMs) demonstrate strong reasoning capabilities in closed-world settings, they struggle in open-ended and dynamic environments. Agentic reasoning marks a paradigm shift by reframing LLMs as autonomous agents that plan, act, and learn through continual interaction. In this survey, we organize agentic reasoning along three complementary dimensions. First, we characterize environmental dynamics through three layers: foundational agentic reasoning, which establishes core single-agent capabilities including planning, tool use, and search in stable environments; self-evolving agentic reasoning, which studies how agents refine these capabilities through feedback, memory, and adaptation; and collective multi-agent reasoning, which extends intelligence to collaborative settings involving coordination, knowledge sharing, and shared goals. Across these layers, we distinguish in-context reasoning, which scales test-time interaction through structured orchestration, from post-training reasoning, which optimizes behaviors via reinforcement learning and supervised fine-tuning. We further review representative agentic reasoning frameworks across real-world applications and benchmarks, including science, robotics, healthcare, autonomous research, and mathematics. This survey synthesizes agentic reasoning methods into a unified roadmap bridging thought and action, and outlines open challenges and future directions, including personalization, long-horizon interaction, world modeling, scalable multi-agent training, and governance for real-world deployment.

Community

Paper submitter

๐ŸŒ Awesome-Agentic-Reasoning GitHub Link:https://github.com/weitianxin/Awesome-Agentic-Reasoning

This survey presents a comprehensive and unified view of agentic reasoning ๐Ÿค–๐Ÿง . It examines how reasoning is integrated into LLM-based agents to drive planning ๐Ÿ—บ๏ธ, search ๐Ÿ”, tool use ๐Ÿ› ๏ธ, memory ๐Ÿง , feedback-driven adaptation ๐Ÿ”„, and multi-agent coordination ๐Ÿ‘ฅ under different dynamics.

The landscape is systematically organized into three settings:
โ€ข Foundational agents: reasoning for planning, search, and tool use
โ€ข Self-evolving agents: learning from feedback and long-term memory
โ€ข Collaborative agents: coordination and communication in multi-agent system reasoning

The survey reviews benchmarks and real-world applications across
๐Ÿงฎ math discovery | ๐ŸŽจ vibe coding | ๐Ÿ”ฌ science | ๐Ÿค– robotics | ๐Ÿฅ healthcare | ๐Ÿง‘โ€๐Ÿ”ฌ autonomous research | ๐ŸŒ web exploration

Based on an analysis of ~800 papers, it highlights key system designs and training paradigms, and outlines open challenges ahead.

Comprehensive coverage, do not miss this

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2601.12538 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2601.12538 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2601.12538 in a Space README.md to link it from this page.

Collections including this paper 3