Probabilistic Verification Framework for RNN Policies in RL
Researchers propose RNN-ProVe, a probabilistic framework for verifying recurrent neural network policies in partially observable reinforcement learning. The method estimates the likelihood of undesired behaviors by using policy-driven sampling to approximate feasible hidden states, deriving statistical error bounds for high-confidence estimates. Experiments on single-agent and cooperative multi-agent tasks demonstrate its effectiveness. The work addresses challenges in verifying history-dependent policies that rely on latent hidden state dynamics, where existing tools often rely on restrictive assumptions or coarse over-approximations.
Key facts
- Proposed framework: RNN-ProVe (RNN Probabilistic Verification)
- Estimates likelihood of undesired behaviors in RNN-based policies
- Uses policy-driven sampling to approximate feasible hidden states
- Derives statistical error bounds for bounded-error, high-confidence estimates
- Applied to partially observable single-agent and cooperative multi-agent tasks
- Addresses limitations of existing RNN verification tools
- Published on arXiv with ID 2605.14758
- Focuses on history-dependent policies induced by recurrent neural networks
Entities
Institutions
- arXiv