ARTFEED — Contemporary Art Intelligence

LAKE: Training-Free Anomaly Detection in Vision-Language Models

publication · 2026-04-30

A recent study on arXiv disputes the belief that vision-language models (VLMs) need additional adapters or memory systems for anomaly detection (AD). The researchers contend that pre-trained models inherently contain anomaly-specific knowledge, which is often dormant and not fully utilized, residing in a limited group of neurons that respond to anomalies. They introduce a framework called latent anomaly knowledge excavation (LAKE), which operates without training and utilizes a small number of normal samples to activate these essential neuronal signals. By targeting these responsive neurons, LAKE creates a highly efficient representation of normality. This research is documented in arXiv:2604.07802v3.

Key facts

  • Paper challenges assumption that VLMs need external adapters for anomaly detection
  • Argues anomaly knowledge is intrinsically embedded but latent
  • Knowledge concentrated in sparse subset of anomaly-sensitive neurons
  • Proposes LAKE, a training-free framework
  • LAKE uses only minimal set of normal samples
  • Constructs compact normality representation
  • Published on arXiv with ID 2604.07802v3
  • Announce type: replace-cross

Entities

Institutions

  • arXiv

Sources