This is a preprint.
Frugal inference for control
- PMID: 39575123
- PMCID: PMC11581108
Frugal inference for control
Abstract
A key challenge in advancing artificial intelligence is achieving the right balance between utility maximization and resource use by both external movement and internal computation. While this trade-off has been studied in fully observable settings, our understanding of resource efficiency in partially observable environments remains limited. Motivated by this challenge, we develop a version of the POMDP framework where the information gained through inference is treated as a resource that must be optimized alongside task performance and motion effort. By solving this problem in environments described by linear-Gaussian dynamics, we uncover fundamental principles of resource efficiency. Our study reveals a phase transition in the inference, switching from a Bayes-optimal approach to one that strategically leaves some uncertainty unresolved. This frugal behavior gives rise to a structured family of equally effective strategies, facilitating adaptation to later objectives and constraints overlooked during the original optimization. We illustrate the applicability of our framework and the generality of the principles we derived using two nonlinear tasks. Overall, this work provides a foundation for a new type of rational computation that both brains and machines could use for effective but resource-efficient control under uncertainty.
Figures
References
-
- Lauri M., Hsu D., Pajarinen J.: Partially observable markov decision processes in robotics: A survey. IEEE Transactions on Robotics 39(1), 21–40 (2022)
-
- Kurniawati H.: Partially observable markov decision processes and robotics. Annual Review of Control, Robotics, and Autonomous Systems 5(1), 253–277 (2022)
-
- Ross S., Pineau J., Chaib-draa B., Kreitmann P.: A bayesian approach for learning and planning in partially observable markov decision processes. Journal of Machine Learning Research 12(5) (2011)
-
- Lim M.H., Becker T.J., Kochenderfer M.J., Tomlin C.J., Sunberg Z.N.: Optimality guarantees for particle belief approximation of pomdps. Journal of Artificial Intelligence Research 77, 1591–1636 (2023)
-
- Watter M., Springenberg J., Boedecker J., Riedmiller M.: Embed to control: A locally linear latent dynamics model for control from raw images. Advances in neural information processing systems 28 (2015)
Publication types
Grants and funding
LinkOut - more resources
Full Text Sources