ArXiv TLDR

Tree-of-Text: A Tree-based Prompting Framework for Table-to-Text Generation in the Sports Domain

🐦 Tweet
2604.26501

Shang-Hsuan Chiang, Tsan-Tsung Yang, An-Zi Yen, Wen-Chih Peng

cs.CLcs.AIcs.HC

TLDR

Tree-of-Text is a tree-based prompting framework that improves LLM table-to-text generation for sports reports by reducing hallucination and boosting efficiency.

Key contributions

  • Introduces Tree-of-Text, a tree-structured prompting framework for table-to-text generation.
  • Guides LLMs through content planning, operation execution on sub-tables, and cohesive report generation.
  • Significantly outperforms existing methods on ShuttleSet+, RotoWire-FG, and MLB datasets.
  • Achieves comparable results with 40% of the time and cost of Chain-of-Table, demonstrating efficiency.

Why it matters

This paper addresses the challenge of LLM hallucination in table-to-text generation for sports reports. Tree-of-Text offers a structured, efficient approach that improves accuracy and reduces computational cost. It provides a promising direction for leveraging LLMs in complex data-to-text tasks.

Original Abstract

Generating sports game reports from structured tables is a complex table-to-text task that demands both precise data interpretation and fluent narrative generation. Traditional model-based approaches require large, annotated datasets, while prompt-based methods using large language models (LLMs) often struggle with hallucination due to weak table comprehension. To overcome these challenges, we propose Tree-of-Text, a tree-structured prompting framework that guides LLMs through a three-stage generation process: (1) Content Planning, where relevant operations and arguments are selected from the input tables; (2) Operation Execution, which breaks down large tables into manageable sub-tables; and (3) Content Generation, where short textual outputs are merged and rewritten into a cohesive report. Experiments show that our method outperforms existing methods on ShuttleSet+, leads in RG and CO metrics on RotoWire-FG, and excels in CS and CO on MLB with roughly 40% of the time and cost of Chain-of-Table. These results demonstrate the effectiveness and efficiency of Tree-of-Text and suggest a promising direction for prompt-based table-to-text generation in the sports domain.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.