PROCO: LLM-Based Safe Reinforcement Learning from Limited Violation Data
A new framework called PROCO has been introduced by researchers, designed for offline safe reinforcement learning using model-based techniques. This approach utilizes large language models (LLMs) to embed natural language insights, tackling the issue of developing policies that adhere to constraints from datasets lacking unsafe samples, a frequent occurrence in critical situations. Traditional methods typically regard all data as equally safe, neglecting safe yet infeasible states that can result in violations. By harnessing LLMs, PROCO enhances safety while avoiding hazardous online interactions. Further information about this framework can be found in arXiv:2605.01356.
Key facts
- PROCO is a model-based offline safe RL framework
- Uses large language models to incorporate natural language knowledge
- Addresses datasets with few or no unsafe samples
- Conventional methods overlook safe-but-infeasible states
- High-stakes scenarios prevent risky trial-and-error
- Framework is described in arXiv:2605.01356
- Aims to learn constraint-satisfying policies without online interaction
- Inspired by knowledge-data integration concept
Entities
Institutions
- arXiv