ArXiv TLDR

Teaching LLMs Human-Like Editing of Inappropriate Argumentation via Reinforcement Learning

🐦 Tweet
2604.12770

Timon Ziegenbein, Maja Stahl, Henning Wachsmuth

cs.CL

TLDR

This paper introduces an RL approach that teaches LLMs to perform human-like, self-contained edits for improving argument appropriateness.

Key contributions

  • Introduces an RL approach for human-like, self-contained editing of arguments.
  • Employs Group Relative Policy Optimization with a multi-component reward function.
  • Optimizes edit-level semantic similarity, fluency, pattern conformity, and argument appropriateness.
  • Achieves state-of-the-art performance in human-like editing and argument appropriateness.

Why it matters

LLMs often struggle with human-like editing, making their suggestions less useful. This work bridges that gap by teaching LLMs to make more natural, meaning-preserving edits. This significantly improves the utility of LLMs for refining human arguments, bringing their editing capabilities closer to human standards.

Original Abstract

Editing human-written text has become a standard use case of large language models (LLMs), for example, to make one's arguments more appropriate for a discussion. Comparing human to LLM-generated edits, however, we observe a mismatch in editing strategies: While LLMs often perform multiple scattered edits and tend to change meaning notably, humans rather encapsulate dependent changes in self-contained, meaning-preserving edits. In this paper, we present a reinforcement learning approach that teaches LLMs human-like editing to improve the appropriateness of arguments. Our approach produces self-contained sentence-level edit suggestions that can be accepted or rejected independently. We train the approach using group relative policy optimization with a multi-component reward function that jointly optimizes edit-level semantic similarity, fluency, and pattern conformity as well as argument-level appropriateness. In automatic and human evaluation, it outperforms competitive baselines and the state of the art in human-like editing, with multi-round editing achieving appropriateness close to full rewriting.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.