ARTFEED — Contemporary Art Intelligence

RiskGate: A Viability-Based Framework for Governing Autonomous AI Agents

publication · 2026-04-29

There's a new preprint on arXiv, numbered 2604.24686, that introduces something called the Agent Viability Framework. This framework is designed to keep an eye on autonomous AI agents while they're functioning, drawing from Aubin's viability theory. It highlights three key features: monitoring, anticipation, and monotonic restriction, which help manage known risks. The authors also put forth the Informational Viability Principle, which sets a limit on hidden dangers using a specific formula. Only when the capacity exceeds this limit, plus a safety buffer, can actions be taken. The implementation, named RiskGate, utilizes various statistical methods and a secure fail-safe system to address safety concerns in AI agents that might change behavior or respond to threats without any code changes.

Key facts

  • arXiv preprint 2604.24686 proposes the Agent Viability Framework for autonomous AI agent governance.
  • The framework is grounded in Aubin's viability theory.
  • Three properties are defined: monitoring (P1), anticipation (P2), and monotonic restriction (P3).
  • The Informational Viability Principle estimates unobserved risk as B̂(x) = U(x) + SB(x) + RG(x).
  • Actions are allowed only when capacity S(x) exceeds B̂(x) by a safety margin.
  • RiskGate instantiates the framework with dedicated statistical estimators.
  • Statistical estimators include KL divergence, segment-vs-rest z-tests, and sequential pattern matching.
  • RiskGate includes a fail-secure monotonic pipeline and a closed-loop Autopilot.

Entities

Institutions

  • arXiv

Sources