ARTFEED — Contemporary Art Intelligence

Curated Synthetic Data Prevents Model Collapse in Generative Retraining

ai-technology · 2026-05-11

A recent theoretical investigation published on arXiv in May 2025 disputes the common belief that recursive retraining of generative models inevitably results in output collapse. The authors demonstrate that by curating synthetic outputs using various reward functions instead of relying on a single, fixed signal, the model can preserve its diversity. This research articulates the dynamics of recursive training amidst differing preferences and establishes convergence towards a stable distribution that spreads probability across competing high-reward areas. This limiting distribution aligns with a weighted Nash bargaining solution, giving a formal interpretation of value aggregation within synthetic retraining processes. The findings challenge earlier assertions that collapse is unavoidable without incorporating real data, thus supporting the use of curated synthetic data in AI training frameworks.

Key facts

  • Recursive retraining of generative models can collapse onto a narrow set of outputs when using a single reward function.
  • The study shows collapse can be mitigated through curation based on multiple reward functions.
  • The model converges to a stable distribution that allocates probability mass across competing high-reward regions.
  • The limiting distribution satisfies a weighted Nash bargaining solution.
  • The paper offers a formal interpretation of value aggregation in synthetic retraining loops.
  • Prior work suggested collapse is unavoidable without adding real data.
  • The research is categorized under Computer Science > Machine Learning.
  • The paper was submitted to arXiv in May 2025.

Entities

Institutions

  • arXiv

Sources