Active Vision-Language Models via Sequential Experimental Design
A new arXiv paper proposes a framework to overcome the perceptual bandwidth bottleneck in Vision-Language Models (VLMs) by framing visual perception as a sequential decision-making process. The authors draw on active vision and information foraging paradigms, formalizing the problem as a sequential Bayesian optimal experimental design (S-BOED) challenge. They develop tractable approximations for continuous gigapixel spaces, balancing spatial coverage and resolution. A training-free inference strategy is presented as a practical instantiation of the S-BOED objective, designed as a flexible template that accommodates arbitrary optimization algorithms. The paper is available at arXiv:2605.01345.
Key facts
- The paper is titled 'Active Reasoning Vision-Language Models via Sequential Experimental Design'.
- It addresses the perceptual bandwidth bottleneck in VLMs.
- The approach is inspired by active vision and information foraging.
- The problem is formalized as a sequential Bayesian optimal experimental design (S-BOED) problem.
- Tractable approximations are derived for continuous gigapixel spaces.
- A training-free inference strategy is presented as a practical instantiation.
- The strategy is a flexible template for agents with multiple vision tools.
- The paper is available on arXiv with ID 2605.01345.
Entities
Institutions
- arXiv