Small Language Models Achieve CPU-Deployable Radiology AI via LoRA Fine-Tuning
A recent study shows that small language models (SLMs) with 3-4 billion parameters can efficiently tackle various radiology tasks when fine-tuned with LoRA, even on standard consumer CPUs. This breakthrough tackles the difficulties of using large language models (LLMs) in resource-limited clinical settings. The research trained the Qwen2.5-3B-Instruct and Qwen3-4B models on 162,000 samples from nine radiology tasks, including RADS classification for ten systems, impression generation, and more, using data from 12 public datasets. Tests on up to 500 withheld samples per task revealed that LoRA fine-tuning led to impressive performance gains: RADS accuracy jumped by 53%, NLI by 60%, and N-staging by 89%. This makes AI in radiology more accessible without needing specialized hardware.
Key facts
- Small language models (SLMs) of 3-4 billion parameters are used.
- LoRA fine-tuning is applied to Qwen2.5-3B-Instruct and Qwen3-4B.
- Training data: 162K samples from 12 public datasets.
- Tasks: 9 radiology tasks including RADS classification, impression generation, temporal comparison, NLI, NER, abnormality detection, N/M staging, radiology Q&A.
- Evaluation: up to 500 held-out test samples per task.
- Performance gains: RADS accuracy +53%, NLI +60%, N-staging +89% over zero-shot baselines.
- Models are deployable on consumer-grade CPUs.
- The two models show complementary performance.
Entities
Institutions
- arXiv