SLAM as a Stochastic Control Problem with Partial Information: Optimal Solutions and Rigorous Approximations
Ilir Gusija, Fady Alajaji, Serdar Yüksel
TLDR
This paper recasts active SLAM as an optimal stochastic control problem, deriving near-optimal solutions for a new POMDP formulation.
Key contributions
- Formulates active SLAM as an optimal stochastic control problem under partial information.
- Introduces a novel exploration stage cost encoding state geometry for information gathering.
- Develops a nonstandard POMDP to derive rigorously justified near-optimal solutions.
- Conducts numerical studies to learn near-optimal policies using standard algorithms.
Why it matters
This work provides a new theoretical framework for active SLAM, treating it as a decision-making problem. Its rigorous approximations and novel cost function could lead to more efficient and robust robotic navigation and mapping systems.
Original Abstract
Simultaneous localization and mapping (SLAM) is a foundational state estimation problem in robotics in which a robot accurately constructs a map of its environment while also localizing itself within this construction. We study the active SLAM problem through the lens of optimal stochastic control, thereby recasting it as a decision-making problem under partial information. After reviewing several commonly studied models, we present a general stochastic control formulation of active SLAM together with a rigorous treatment of motion, sensing, and map representation. We introduce a new exploration stage cost that encodes the geometry of the state when evaluating information-gathering actions. This formulation, constructed as a nonstandard partially observable Markov decision process (POMDP), is then analyzed to derive rigorously justified approximate solutions that are near-optimal. To enable this analysis, the associated regularity conditions are studied under general assumptions that apply to a wide range of robotics applications. For a particular case, we conduct an extensive numerical study in which standard learning algorithms are used to learn near-optimal policies.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.