ARTFEED — Contemporary Art Intelligence

Zero-Shot ODD Perception with Vision-Language Models

other · 2026-05-11

A new arXiv paper (2605.07649) explores using vision-language models (VLMs) as zero-shot sensors for Operational Design Domain (ODD) perception in autonomous systems, particularly Automated Driving Systems (ADS). The research addresses the need for adaptable perception that can handle evolving ODD definitions without task-specific training data. VLMs integrate visual recognition and language reasoning, making them suitable for this role. The study contributes to safety regulation compliance in autonomous driving.

Key facts

  • Paper ID: arXiv:2605.07649
  • Focus on zero-shot ODD perception using VLMs
  • Relevant for Automated Driving Systems (ADS)
  • VLMs function without task-specific training data
  • Addresses safety regulation adherence
  • ODD defines conditions for autonomous agent function
  • Research aims to translate autonomous systems into practical use cases
  • VLMs act as adaptable 'ODD sensors'

Entities

Institutions

  • arXiv

Sources