ARTFEED — Contemporary Art Intelligence

LLM Agent Adaptation via Semantic and Episodic Memory

ai-technology · 2026-05-04

A new memory-augmented framework enables large language model (LLM) agents to learn classification tasks from labeled examples without parameter updates. The approach stores instance-level critiques in episodic memory and distills them into reusable task-level guidance in semantic memory. Across diverse tasks and models, the best self-critique strategy using both memory types achieved an average improvement of 8.1 percentage points over zero-shot baselines and 4.6 points over a RAG-based baseline relying only on labels. Performance varied substantially across models and domains, leading the researchers to introduce 'suggestibility' as a metric to explain this variation. The paper is available on arXiv under identifier 2510.19897.

Key facts

  • Framework uses episodic and semantic memory for LLM agent adaptation without parameter updates.
  • Best self-critique strategy improves by 8.1 percentage points over zero-shot baseline.
  • Improves by 4.6 percentage points over RAG-based baseline.
  • Performance varies across models and domains.
  • Introduces 'suggestibility' metric to explain variation.
  • Paper available on arXiv: 2510.19897.

Entities

Institutions

  • arXiv

Sources