Robust Quantile-based Implicit Quantile Networks for Distributional RL
A new paper on arXiv proposes Robust Quantile-based Implicit Quantile Networks (RQIQN), a method to improve distributional reinforcement learning. The approach addresses distortions in quantile-based return distribution estimates by applying Wasserstein distributionally robust optimization to each quantile slot. This yields a closed-form correction to the Bellman target that preserves the risk-neutral average while preventing distributional collapse. The work is purely theoretical and algorithmic, with no direct application to art or culture.
Key facts
- arXiv:2605.08182v1
- Announce Type: cross
- Proposes Robust Quantile-based Implicit Quantile Networks (RQIQN)
- Uses Wasserstein distributionally robust enhancement
- Reinterprets IQN loss as local empirical quantile estimation problems
- Derives fraction-dependent correction to Bellman target
- Correction preserves risk-neutral quantile average
- Correction prevents distributional collapse
Entities
Institutions
- arXiv