Dissociating model architectures from inference computations
- PMID: 40673431
- DOI: 10.1080/17588928.2025.2532604
Dissociating model architectures from inference computations
Abstract
Parr et al., 2025 examines how auto-regressive and deep temporal models differ in their treatment of non-Markovian sequence modelling. Building on this, we highlight the need for dissociating model architectures-i.e., how the predictive distribution factorises-from the computations invoked at inference. We demonstrate that deep temporal computations are mimicked by autoregressive models by structuring context access during iterative inference. Using a transformer trained on next-token prediction, we show that inducing hierarchical temporal factorisation during iterative inference maintains predictive capacity while instantiating fewer computations. This emphasises that processes for constructing and refining predictions are not necessarily bound to their underlying model architectures.
Keywords: Deep temporal structures; language models; structured context access; transformers.
MeSH terms
LinkOut - more resources
Full Text Sources