ArXiv TLDR

Rhetorical Questions in LLM Representations: A Linear Probing Study

🐦 Tweet
2604.14128

Louie Hong Yao, Vishesh Anand, Yuan Zhuang, Tianyu Jiang

cs.CLcs.AIcs.LG

TLDR

LLMs encode rhetorical questions via multiple linear directions, not a single shared one, with signals emerging early in representations.

Key contributions

  • LLMs represent rhetorical questions distinctly from information-seeking ones, with signals emerging early.
  • Linear probes achieve AUROC 0.7-0.8 in detecting rhetorical questions, even across different datasets.
  • Cross-dataset transfer doesn't imply a shared representation; probes capture distinct rhetorical cues.
  • Divergences show some probes capture discourse stance, while others focus on syntax-driven acts.

Why it matters

This paper reveals the nuanced way LLMs encode rhetorical questions, showing they use multiple internal representations. Understanding these distinct encodings is crucial for improving LLM's ability to grasp subtle human communication and persuasive language. It highlights the complexity of linguistic phenomena within LLMs.

Original Abstract

Rhetorical questions are asked not to seek information but to persuade or signal stance. How large language models internally represent them remains unclear. We analyze rhetorical questions in LLM representations using linear probes on two social-media datasets with different discourse contexts, and find that rhetorical signals emerge early and are most stably captured by last-token representations. Rhetorical questions are linearly separable from information-seeking questions within datasets, and remain detectable under cross-dataset transfer, reaching AUROC around 0.7-0.8. However, we demonstrate that transferability does not simply imply a shared representation. Probes trained on different datasets produce different rankings when applied to the same target corpus, with overlap among the top-ranked instances often below 0.2. Qualitative analysis shows that these divergences correspond to distinct rhetorical phenomena: some probes capture discourse-level rhetorical stance embedded in extended argumentation, while others emphasize localized, syntax-driven interrogative acts. Together, these findings suggest that rhetorical questions in LLM representations are encoded by multiple linear directions emphasizing different cues, rather than a single shared direction.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.