Position: LLMs Need a Bayesian Meta-Reasoning Framework for More Robust and Generalizable Reasoning

Hanqi Yan*, Linhai Zhang, Jiazheng Li, Zhenyi Shen, Yulan He*

*Corresponding author for this work

Research output: Contribution to conference typesPaperpeer-review

262 Downloads (Pure)

Abstract

Large language models (LLMs) excel in many reasoning tasks but continue to face significant challenges, such as lack of robustness in reasoning, struggling with cross-task generalization, and inefficiencies in scaling up reasoning capabilities. Current training paradigms, including next-token prediction and reinforcement learning from human feedback, often fall short in adaptability to diverse reasoning tasks. Existing approaches, such as prompt optimization and iterative output refinement, offer performance improvement, but can be inefficient and lack effective generalization. To overcome these limitations, this position paper argues for a transformative shift in how LLMs approach reasoning. Drawing inspiration from cognitive science, particularly meta-reasoning theories such as Dual-Process Theory and Metacognitive Reasoning, we propose a Bayesian meta-reasoning framework for LLMs. Our approach integrates self-awareness, monitoring, evaluation, regulation, and meta-reflection, to enhance LLMs’ ability to refine reasoning strategies and generalize across tasks. We revisit existing LLM reasoning methods, identify key challenges, and suggest directions for future research.
Original languageEnglish
Publication statusAccepted/In press - 1 May 2025
Event2025 International Conference on Machine Learning: ICML25 -
Duration: 13 Jul 2025 → …

Conference

Conference2025 International Conference on Machine Learning
Period13/07/2025 → …

Keywords

  • meta-reasoning
  • Large Language Models (LLMs)
  • Position Paper

Fingerprint

Dive into the research topics of 'Position: LLMs Need a Bayesian Meta-Reasoning Framework for More Robust and Generalizable Reasoning'. Together they form a unique fingerprint.

Cite this