ArXiv TLDR

Qwen Goes Brrr: Off-the-Shelf RAG for Ukrainian Multi-Domain Document Understanding

🐦 Tweet
2605.10296

Anton Bazdyrev, Ivan Bashtovyi, Ivan Havlytskyi, Oleksandr Kharytonov, Artur Khodakovskyi

cs.CLcs.AIcs.IRcs.LG

TLDR

A Qwen-based RAG system achieves high accuracy in Ukrainian multi-domain document understanding using contextual chunking and question-aware reranking.

Key contributions

  • Proposes a retrieval-augmented pipeline for Ukrainian multi-domain document understanding.
  • Utilizes contextual chunking of PDFs to preserve document structure.
  • Employs question-aware dense retrieval and reranking conditioned on questions and answer options.
  • Achieved 0.9674 answer accuracy using top-2 reranked passages on a held-out split.

Why it matters

This paper introduces an effective Qwen-based RAG pipeline for challenging Ukrainian multi-domain document understanding. It demonstrates that preserving document structure and integrating answer-space awareness into relevance estimation are key for robust performance, achieving strong results under strict constraints.

Original Abstract

We participated in the Fifth UNLP shared task on multi-domain document understanding, where systems must answer Ukrainian multiple-choice questions from PDF collections and localize the supporting document and page. We propose a retrieval-augmented pipeline built around three ideas: contextual chunking of PDFs, question-aware dense retrieval and reranking conditioned on both the question and answer options, and constrained answer generation from a small set of reranked passages. Our final system uses Qwen3-Embedding-8B for retrieval, a fine-tuned Qwen3-Reranker-8B for passage ranking, and Qwen3-32B for answer selection. On a held-out split, reranking improves Recall@1 from 0.6957 to 0.7935, while using the top-2 reranked passages raises answer accuracy from 0.9348 to 0.9674. Our best leaderboard run reached 0.9452 on the public leaderboard and 0.9598 on the private leaderboard. Our results suggest that, under strict code-competition constraints, preserving document structure and making relevance estimation aware of the answer space are more effective than adding complex downstream heuristics.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.