Meta-Tool: Hypernetwork LoRA Fails to Improve Few-Shot Tool Use in 3B LLM
A new study from arXiv (2604.20148) finds that hypernetwork-based LoRA adaptation provides no measurable improvement over few-shot prompting for small language models in tool-use tasks. Using a Llama-3.2-3B-Instruct backbone, researchers evaluated four adaptation mechanisms—few-shot prompting, documentation encoding, hypernetwork-generated LoRA weights, and value-guided beam search—across four benchmarks: Gorilla APIBench, Spider 2.0, WebArena, and InterCode. The 227.8M-parameter hypernetwork added 0% performance gain, while few-shot examples contributed +21.5% and documentation +5.0%. The negative result suggests that careful prompting alone suffices for 3B-scale models.
Key facts
- Meta-Tool is a controlled empirical study comparing hypernetwork-based LoRA adaptation against few-shot prompting.
- Uses Llama-3.2-3B-Instruct backbone.
- Evaluated on Gorilla APIBench, Spider 2.0, WebArena, and InterCode.
- Hypernetwork has 227.8M parameters.
- Few-shot examples contribute +21.5% to performance.
- Documentation contributes +5.0%.
- Hypernetwork adds 0% improvement.
- Study concludes that few-shot prompting alone is sufficient for 3B models.
Entities
Institutions
- arXiv