ARTFEED — Contemporary Art Intelligence

Probabilistic Verification Framework for RNN Policies in RL

ai-technology · 2026-05-16

Researchers propose RNN-ProVe, a probabilistic framework for verifying recurrent neural network policies in partially observable reinforcement learning. The method estimates the likelihood of undesired behaviors by using policy-driven sampling to approximate feasible hidden states, deriving statistical error bounds for high-confidence estimates. Experiments on single-agent and cooperative multi-agent tasks demonstrate its effectiveness. The work addresses challenges in verifying history-dependent policies that rely on latent hidden state dynamics, where existing tools often rely on restrictive assumptions or coarse over-approximations.

Key facts

  • Proposed framework: RNN-ProVe (RNN Probabilistic Verification)
  • Estimates likelihood of undesired behaviors in RNN-based policies
  • Uses policy-driven sampling to approximate feasible hidden states
  • Derives statistical error bounds for bounded-error, high-confidence estimates
  • Applied to partially observable single-agent and cooperative multi-agent tasks
  • Addresses limitations of existing RNN verification tools
  • Published on arXiv with ID 2605.14758
  • Focuses on history-dependent policies induced by recurrent neural networks

Entities

Institutions

  • arXiv

Sources