Breaking the Capability Ceiling of LLM Post-Training by Reintroducing Markov States

By: Daniel White, Olivia Green, Sophia Black, Liam Grey

Published: 2026-03-20

View on arXiv →
#cs.AI

Abstract

This paper proposes a novel approach to enhance the capabilities of large language models during post-training by reintroducing Markov states. It argues that current post-training paradigms often hit a "capability ceiling" due to limitations in capturing complex sequential dependencies. By integrating Markov states, the model can better learn and generalize across diverse tasks, leading to significant improvements in reasoning, planning, and long-context understanding, unlocking new potentials for LLM applications.

FEEDBACK

Projects

No projects yet

Breaking the Capability Ceiling of LLM Post-Training by Reintroducing Markov States | ArXiv Intelligence