ArXiv TLDR

An Answer is just the Start: Related Insight Generation for Open-Ended Document-Grounded QA

🐦 Tweet
2604.19685

Saransh Sharma, Pritika Ramu, Aparna Garimella, Koyel Mukherjee

cs.CL

TLDR

This paper introduces a new task and dataset for generating related insights from documents to enhance open-ended QA, proposing InsightGen, a two-stage LLM-based approach.

Key contributions

  • Introduces "document-grounded related insight generation," a new task to extend initial QA answers.
  • Curates SCOpE-QA, a dataset of 3,000 open-ended questions across 20 scientific collections.
  • Proposes InsightGen, a two-stage approach using thematic clustering and LLMs to generate diverse insights.
  • Establishes a strong baseline, demonstrating InsightGen's ability to produce useful and actionable insights.

Why it matters

Open-ended QA needs more than single answers; users refine responses. This work addresses this by enabling AI to generate related insights, improving user interaction and the overall QA experience. It pushes AI beyond simple factual retrieval.

Original Abstract

Answering open-ended questions remains challenging for AI systems because it requires synthesis, judgment, and exploration beyond factual retrieval, and users often refine answers through multiple iterations rather than accepting a single response. Existing QA benchmarks do not explicitly support this refinement process. To address this gap, we introduce a new task, document-grounded related insight generation, where the goal is to generate additional insights from a document collection that help improve, extend, or rethink an initial answer to an open-ended question, ultimately supporting richer user interaction and a better overall question answering experience. We curate and release SCOpE-QA (Scientific Collections for Open-Ended QA), a dataset of 3,000 open-ended questions across 20 research collections. We present InsightGen, a two-stage approach that first constructs a thematic representation of the document collection using clustering, and then selects related context based on neighborhood selection from the thematic graph to generate diverse and relevant insights using LLMs. Extensive evaluation on 3,000 questions using two generation models and two evaluation settings shows that InsightGen consistently produces useful, relevant, and actionable insights, establishing a strong baseline for this new task.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.