On Information Self-Locking in Reinforcement Learning for Active Reasoning of LLM agents

By: Deyu Zou, Yongqiang Chen, Fan Feng, Mufei Li, Pan Li, Yu Gong, James Cheng

Published: 2026-03-13

View on arXiv →
#cs.AI

Abstract

This paper explores the phenomenon of "information self-locking" in reinforcement learning for active reasoning in Large Language Model (LLM) agents. It investigates how LLM agents might get stuck in suboptimal reasoning loops and proposes methods to overcome these limitations for improved active reasoning.

FEEDBACK

Projects

No projects yet

On Information Self-Locking in Reinforcement Learning for Active Reasoning of LLM agents | ArXiv Intelligence