ArXiv TLDR

Governing Reflective Human-AI Collaboration: A Framework for Epistemic Scaffolding and Traceable Reasoning

🐦 Tweet
2604.14898

Rikard Rosenbacke, Carl Rosenbacke, Victor Rosenbacke, Martin McKee

cs.AIcs.CYcs.HC

TLDR

This paper proposes a framework for human-AI collaboration that treats reasoning as a shared process, enabling auditable and accountable AI use.

Key contributions

  • Proposes a framework treating human-AI reasoning as a relational, distributed process, not internal to either.
  • Introduces "The Architect's Pen" method, using models as external tools for structured human reflection.
  • Defines a reasoning loop (human abstraction -> model articulation -> human reflection) for collaborative intelligence.
  • Enables auditable reasoning traces and aligns with AI governance standards like the EU AI Act.

Why it matters

This paper provides a crucial framework for governing human-AI collaboration, shifting focus to a shared reasoning process. It offers a practical path for transparent, accountable AI use, aligning with emerging standards like the EU AI Act, without new model architectures.

Original Abstract

Large language models have advanced rapidly, from pattern recognition to emerging forms of reasoning, yet they remain confined to linguistic simulation rather than grounded understanding. They can produce fluent outputs that resemble reflection, but lack temporal continuity, causal feedback, and anchoring in real-world interaction. This paper proposes a complementary approach in which reasoning is treated as a relational process distributed between human and model rather than an internal capability of either. Building on recent work on "System-2" learning, we relocate reflective reasoning to the interaction layer. Instead of engineering reasoning solely within models, we frame it as a cognitive protocol that can be structured, measured, and governed using existing systems. This perspective emphasizes collaborative intelligence, combining human judgment and contextual understanding with machine speed, memory, and associative capacity. We introduce "The Architect's Pen" as a practical method. Like an architect who thinks through drawing, the human uses the model as an external medium for structured reflection. By embedding phases of articulation, critique, and revision into human-AI interaction, the dialogue itself becomes a reasoning loop: human abstraction -> model articulation -> human reflection. This reframes the question from whether the model can think to whether the human-AI system can reason. The framework enables auditable reasoning traces and supports alignment with emerging governance standards, including the EU AI Act and ISO/IEC 42001. It provides a practical path toward more transparent, controllable, and accountable AI use without requiring new model architectures.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.