ArXiv TLDR

GA3T: A Ground-Aerial Terrain Traversability Dataset for Heterogeneous Robot Teams in Unstructured Environments

🐦 Tweet
2605.06478

Siwei Cai, Knut Peterson, Quan Tran, Christian Ricks, Dhanush Parthasarathy + 5 more

cs.RO

TLDR

GA3T is a new real-world dataset for heterogeneous air-ground robot teams, enabling collaborative perception in diverse unstructured environments.

Key contributions

  • Introduces GA3T, a real-world dataset for air-ground robot collaborative perception.
  • Features multi-modal data from UGV (LiDAR, stereo) and UAV (RGB, thermal) in diverse terrains.
  • Includes over 13,000 synchronized frames with SAM-3 segmentation and 8,000+ manual labels.
  • Supports research in cross-view fusion, traversability, and collaborative scene understanding.

Why it matters

This paper introduces GA3T, a crucial dataset addressing the lack of real-world data for heterogeneous air-ground robot teams. It enables advancements in collaborative perception, traversability estimation, and scene understanding in complex off-road environments. Its unique early-spring collection also supports occlusion-aware perception research.

Original Abstract

Heterogeneous air-ground robot teams combine complementary sensing modalities, mobility characteristics, and spatial viewpoints that can significantly enhance perception in complex outdoor environments. However, progress in multi-robot collaborative perception has been constrained by the lack of real-world datasets featuring overlapping multi-modal observations from platforms operating in unstructured terrain. We present GA3T (Ground-Aerial Team for Terrain Traversal), a real-world multi-robot collaborative perception dataset collected using a Clearpath Husky UGV and an Autel EVO~II UAV across diverse unstructured environments, including forest trails, rocky paths, muddy terrain, snow piles, and grass-covered fields. The ground platform provides 3D LiDAR, stereo camera, IMU, and GPS data, while the aerial platform contributes RGB imagery, thermal/infrared observations, and GPS from a complementary overhead viewpoint, allowing for rich cross-modal and cross-view perception. The dataset is collected in 4 unique environments, with over 13,000 synchronized frames across approximately 29 minutes of operation, and includes both SAM~3-based zero-shot segmentation and over 8,000 manually labeled images. A unique aspect of the dataset is its early-spring collection period, during which sparse tree canopies allow the aerial robot to partially observe the ground robot and terrain through the trees, allowing for occlusion-aware collaborative perception. Unlike prior multi-robot datasets that focus on SLAM or simulated cooperative driving, GA3T is specifically designed to support research on cross-view perception, air-ground viewpoint fusion, traversability estimation, and collaborative scene understanding in real off-road environments.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.