ARTFEED — Contemporary Art Intelligence

PaperGym: Cross-Domain Retrieval Boosts LLM Ideation but Not Beyond Random Seeds

publication · 2026-05-13

A recent investigation published on arXiv (ID 2605.11532) examines whether large language model (LLM) ideation systems gain advantages from specific cross-domain retrieval or merely from varied exposure. The authors present PaperGym, which consists of three phases: the initial tool-enhanced seed extraction using read, grep, and bash in controlled paper settings; the second phase involves cross-domain seed retrieval through paraphrasing across seven machine learning sectors; and finally, method synthesis from the retrieved seeds evaluated by rubric-based judges. While tool-enhanced extraction increased specificity and paraphrase retrieval expanded domain coverage, cross-domain retrieval did not significantly outperform a random diverse-seed control, suggesting that LLM ideation systems may not need targeted cross-domain retrieval for novelty.

Key facts

  • PaperGym is a three-stage pipeline for LLM ideation.
  • Stages: tool-augmented seed extraction, cross-domain retrieval, method synthesis.
  • Cross-domain retrieval uses paraphrasing across seven ML domains.
  • Tool-augmented extraction improved specificity.
  • Cross-domain retrieval had more pairwise novelty wins than no-retrieval and same-domain baselines.
  • No significant difference from random diverse-seed control.
  • Study published on arXiv with ID 2605.11532.
  • Research questions whether targeted cross-domain retrieval is necessary.

Entities

Institutions

  • arXiv

Sources