SLMs Tested for Fine-Grained Emotion Preservation in Translation
A recent investigation analyzes three Small Language Models (SLMs)—EuroLLM, Aya Expanse, and Gemma—focusing on their capacity to maintain nuanced emotions during backtranslation in five European languages: German, French, Spanish, Italian, and Polish. Researchers utilize the GoEmotions dataset, which comprises Reddit comments categorized into 28 emotions, to evaluate emotional fidelity in addition to semantic accuracy. Furthermore, the study examines the effectiveness of emotion-aware prompting and contrasts ModernBERT with BERT for emotion classification in machine translation assessments.
Key facts
- Three SLMs evaluated: EuroLLM, Aya Expanse, Gemma
- Languages: German, French, Spanish, Italian, Polish
- Dataset: GoEmotions (28 emotion categories from Reddit)
- Focus on fine-grained emotion preservation in backtranslation
- Investigates inherent SLM capability, emotion-aware prompting, and ModernBERT performance
- ModernBERT compared to BERT for emotion classification in MT evaluation
Entities
Institutions
- arXiv