Rethinking State Tracking in Recurrent Models Through Error Control Dynamics
Abstract
Affine recurrent networks cannot correct errors in state tracking once state representations are preserved, leading to finite horizon solutions governed by accumulated error rather than robust tracking.
The theory of state tracking in recurrent architectures has predominantly focused on expressive capacity: whether a fixed architecture can theoretically realize a set of symbolic transition rules. We argue that equally important is error control, the dynamics governing hidden-state drift along the directions that distinguish symbolic states. We prove that affine recurrent networks, a class of models encompassing State-Space Models and Linear Attention, cannot correct errors along state-separating subspaces once they preserve state representations. Consequently, practical affine trackers do not learn robust state tracking; rather, they learn finite horizon solutions governed by accumulated state-relevant error. We characterize the mechanics of this failure, showing that tracking remains readable only while the accumulating within-class spread remains small relative to the initial between-class separation. We demonstrate empirically on group state-tracking tasks that this breakdown is predictable: tracking collapses when the distinguishability ratio crosses the readability threshold of the trained decoder. Across trained models, the point of this crossing predicts the horizon at which downstream accuracy fails. These results establish that robust state tracking is determined not only by an architecture's theoretical expressivity but crucially by its error control.
Community
This work argues that recurrent state tracking should be understood not only through expressive capacity, but also through error control. We show that affine recurrent architectures, including State-Space Models and Linear Attention, cannot correct perturbations along state-separating subspaces once they preserve symbolic state representations. As a result, practical affine trackers learn finite-horizon solutions rather than robust symbolic dynamics: tracking remains readable only while accumulated within-class drift stays small relative to between-class separation. We formalize this mechanism through a distinguishability ratio and show empirically on group state-tracking tasks that downstream failure occurs when this ratio crosses the trained decoder’s readability threshold. Across trained models, the predicted crossing point closely matches the horizon at which accuracy collapses. These results suggest that robust state tracking depends not only on whether an architecture can express transition rules, but on whether its dynamics can actively control state-relevant errors.
Get this paper in your agent:
hf papers read 2605.07755 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper