ArXiv TLDR

Sentiment analysis for software engineering: How far can zero-shot learning (ZSL) go?

🐦 Tweet
2604.13826

Reem Alfayez, Manal Binkhonain

cs.SEcs.AI

TLDR

Zero-shot learning (ZSL) shows promise for sentiment analysis in software engineering, matching fine-tuned models and reducing the need for costly annotated datasets.

Key contributions

  • Evaluated various ZSL techniques (embedding, NLI, TARS, generative) for sentiment analysis in software engineering.
  • Found ZSL, especially with expert labels, achieves macro-F1 scores comparable to fine-tuned transformer models.
  • Identified subjectivity in annotation and "polar facts" as primary causes of ZSL misclassifications.
  • Demonstrates ZSL's potential to overcome the scarcity of annotated datasets in SE sentiment analysis.

Why it matters

This paper addresses the critical challenge of data scarcity in software engineering sentiment analysis. By demonstrating ZSL's effectiveness, it offers a viable path to developing specialized tools without extensive manual annotation. This could significantly accelerate research and application in the domain.

Original Abstract

Sentiment analysis in software engineering focuses on understanding emotions expressed in software artifacts. Previous research highlighted the limitations of applying general off-the-shelf sentiment analysis tools within the software engineering domain and indicated the need for specialized tools tailored to various software engineering contexts. The development of such tools heavily relies on supervised machine learning techniques that necessitate annotated datasets. Acquiring such datasets is a substantial challenge, as it requires domain-specific expertise and significant effort. Objective: This study explores the potential of ZSL to address the scarcity of annotated datasets in sentiment analysis within software engineering Method:} We conducted an empirical experiment to evaluate the performance of various ZSL techniques, including embedding-based, NLI-based, TARS-based, and generative-based ZSL techniques. We assessed the performance of these techniques under different labels setups to examine the impact of label configurations. Additionally, we compared the results of the ZSL techniques with state-of-the-art fine-tuned transformer-based models. Finally, we performed an error analysis to identify the primary causes of misclassifications. Results: Our findings demonstrate that ZSL techniques, particularly those combining expert-curated labels with embedding-based or generative-based models, can achieve macro-F1 scores comparable to fine-tuned transformer-based models. The error analysis revealed that subjectivity in annotation and polar facts are the main contributors to ZSL misclassifications. Conclusion: This study demonstrates the potential of ZSL for sentiment analysis in software engineering. ZSL can provide a solution to the challenge of annotated dataset scarcity by reducing reliance on annotated dataset.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.