ArXiv TLDR

Zero-Shot to Full-Resource: Cross-lingual Transfer Strategies for Aspect-Based Sentiment Analysis

🐦 Tweet
2604.26619

Jakob Fehle, Nils Constantin Hellwig, Udo Kruschwitz, Christian Wolff

cs.CL

TLDR

This paper evaluates cross-lingual transfer strategies for Aspect-Based Sentiment Analysis across seven languages, finding LLMs excel in complex tasks.

Key contributions

  • Evaluates state-of-the-art ABSA across 7 languages and 4 subtasks.
  • Compares transformer architectures (LLMs, encoders) under zero-shot, data-only, and full-resource settings.
  • Fine-tuned LLMs excel in complex generative tasks; cross-lingual training boosts their transfer.
  • Smaller models benefit most from code-switching; two new German ABSA datasets introduced.

Why it matters

This work systematically addresses the English-centric bias in ABSA by evaluating diverse models and transfer strategies across multiple languages. It provides crucial insights into optimizing cross-lingual performance for different architectures, fostering broader multilingual ABSA research.

Original Abstract

Aspect-based Sentiment Analysis (ABSA) extracts fine-grained opinions toward specific aspects within text but remains largely English-focused despite major advances in transformer-based and instruction-tuned models. This work presents a multilingual evaluation of state-of-the-art ABSA approaches across seven languages (English, German, French, Dutch, Russian, Spanish, and Czech) and four subtasks (ACD, ACSA, TASD, ASQP). We systematically compare different transformer architectures under zero-resource, data-only, and full-resource settings, using cross-lingual transfer, code-switching and machine translation. Fine-tuned Large Language Models (LLMs) achieve the highest overall scores, particularly in complex generative tasks, while few-shot counterparts approach this performance in simpler setups, where smaller encoder models also remain competitive. Cross-lingual training on multiple non-target languages yields the strongest transfer for fine-tuned LLMs, while smaller encoder or seq-to-seq models benefit most from code-switching, highlighting architecture-specific strategies for multilingual ABSA. We further contribute two new German datasets, an adapted GERestaurant and the first German ASQP dataset (GERest), to encourage multilingual ABSA research beyond English.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.