ARTFEED — Contemporary Art Intelligence

Fine-Tuning Compact LLMs for Controllable Children's Story Generation

ai-technology · 2026-05-14

Researchers have developed a method to generate children's English reading stories using compact 8B-parameter large language models (LLMs) fine-tuned on expert-designed curricula. The study, published on arXiv, addresses two key issues: stories generated by LLMs are often too difficult for children, and the high operational cost of large models limits their educational use. By fine-tuning three 8B LLMs on stories from GPT-4o and Llama 3.3 70B, the team achieved better control over reading difficulty and error patterns compared to zero-shot generation from larger models. The approach prioritizes controllability over scale, allowing educators to target specific reading levels with affordable models. Evaluation results showed that fine-tuned 8B models outperformed larger zero-shot models on difficulty-related metrics.

Key facts

  • Study uses 8B-parameter LLMs for children's story generation
  • Fine-tuning based on expert-designed children's reading curriculum
  • Stories from GPT-4o and Llama 3.3 70B used for training
  • Method prioritizes controllability over scale
  • Educators can target reading levels and error patterns
  • Fine-tuned 8B models outperformed larger zero-shot models on difficulty metrics
  • Published on arXiv with ID 2605.13709
  • Aims to reduce operational cost for educational settings

Entities

Institutions

  • arXiv
  • GPT-4o
  • Llama 3.3 70B

Sources