Participatory provenance as representational auditing for AI-mediated public consultation
TLDR
Introduces participatory provenance to audit AI summaries for faithful public input representation in policy consultations.
Key contributions
- Defines participatory provenance using optimal transport, causal inference, and semantic analysis.
- Analyzes Canada's 2025-2026 AI Strategy consultation with 5,253 respondents.
- Finds official summaries exclude 15-17% of participants, especially dissenting voices.
- Provides an open-source tool for policymakers to audit and improve AI-generated summaries.
Why it matters
This paper addresses the accountability gap in AI summarization of public input by ensuring summaries truly reflect diverse voices. It enables transparent, human-in-the-loop oversight, crucial for fair policy-making.
Original Abstract
Artificial intelligence is increasingly deployed to synthesize large-scale public input in policy consultations and participatory processes. Yet no formal framework exists for auditing whether these summaries faithfully represent the source population, an accountability gap that existing approaches to AI explainability, grounding and hallucination detection do not address because they focus on output quality rather than input fidelity. Here, participatory provenance is introduced: a measurement framework grounded in optimal transport theory, causal inference and semantic analysis that tracks how individual public submissions are transformed, filtered or lost through AI-mediated summarization. Applied to Canada's 2025-2026 national AI Strategy consultation ($n = 5{,}253$ respondents across two independent policy topics), the framework reveals that both official government summaries underperform a random-participant baseline ($-9.1\%$ and $-8.0\%$ coverage degradation), with $16.9\%$ and $15.3\%$ of participants effectively excluded. Exclusion concentrates in clusters expressing dissent, scepticism and critique of AI ($33$-$88\%$ exclusion rates). Brevity, semantic isolation and rhetorical register independently predict representational outcome. An accompanying open-source interactive tool, the Co-creation Provenance Lab, enables policymakers to audit and iteratively improve summaries, establishing genuine human-in-the-loop oversight at scale.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.