ArXiv TLDR

Say the Mission, Execute the Swarm: Agent-Enhanced LLM Reasoning in the Web-of-Drones

🐦 Tweet
2605.03788

Andrea Iannoli, Lorenzo Gigli, Luca Sciullo, Angelo Trotta, Marco Di Felice

cs.AIcs.NIcs.RO

TLDR

An agent-enhanced LLM framework enables natural language control of UAV swarms using Web-of-Things standards for grounded, real-time execution.

Key contributions

  • Introduces an agent-enhanced LLM framework for natural language control of UAV swarms.
  • Leverages W3C Web of Things (WoT) standards to abstract drones, sensors, and services.
  • Enables grounded, real-time interaction and safe actuation without relying on code generation.
  • Highlights the need for explicit grounding and execution support for reliable LLM-based swarm tasks.

Why it matters

This paper addresses the challenge of applying LLMs to real-time UAV swarm management, which struggles with heterogeneous interfaces and limited grounding. It offers a robust framework that enables natural language control and safer, more reliable execution. This is crucial for advancing LLM-driven autonomous systems in complex cyber-physical environments.

Original Abstract

Large Language Models (LLMs) are increasingly explored as high-level reasoning engines for cyber-physical systems, yet their application to real-time UAV swarm management remains challenging due to heterogeneous interfaces, limited grounding, and the need for long-running closed-loop execution. This paper presents a mission-agnostic, agent-enhanced LLM framework for UAV swarm control, where users express mission objectives in natural language and the system autonomously executes them through grounded, real-time interactions. The proposed architecture combines an LLM-based Agent Core with a Model Context Protocol (MCP) gateway and a Web-of-Drones abstraction based on W3C Web of Things (WoT) standards. By exposing drones, sensors, and services as standardized WoT Things, the framework enables structured tool-based interaction, continuous state observation, and safe actuation without relying on code generation. We evaluate the framework using ArduPilot-based simulation across four swarm missions and six state-of-the-art LLMs. Results show that, despite strong reasoning abilities, current general-purpose LLMs still struggle to achieve reliable execution - even for simple swarm tasks - when operating without explicit grounding and execution support. Task-specific planning tools and runtime guardrails substantially improve robustness, while token consumption alone is not indicative of execution quality or reliability.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.