ARTFEED — Contemporary Art Intelligence

Meta-Tool: Hypernetwork LoRA Fails to Improve Few-Shot Tool Use in 3B LLM

ai-technology · 2026-04-24

A new study from arXiv (2604.20148) finds that hypernetwork-based LoRA adaptation provides no measurable improvement over few-shot prompting for small language models in tool-use tasks. Using a Llama-3.2-3B-Instruct backbone, researchers evaluated four adaptation mechanisms—few-shot prompting, documentation encoding, hypernetwork-generated LoRA weights, and value-guided beam search—across four benchmarks: Gorilla APIBench, Spider 2.0, WebArena, and InterCode. The 227.8M-parameter hypernetwork added 0% performance gain, while few-shot examples contributed +21.5% and documentation +5.0%. The negative result suggests that careful prompting alone suffices for 3B-scale models.

Key facts

  • Meta-Tool is a controlled empirical study comparing hypernetwork-based LoRA adaptation against few-shot prompting.
  • Uses Llama-3.2-3B-Instruct backbone.
  • Evaluated on Gorilla APIBench, Spider 2.0, WebArena, and InterCode.
  • Hypernetwork has 227.8M parameters.
  • Few-shot examples contribute +21.5% to performance.
  • Documentation contributes +5.0%.
  • Hypernetwork adds 0% improvement.
  • Study concludes that few-shot prompting alone is sufficient for 3B models.

Entities

Institutions

  • arXiv

Sources