ArXiv TLDR

NORACL: Neurogenesis for Oracle-free Resource-Adaptive Continual Learning

🐦 Tweet
2604.27031

Karthik Charan Raghunathan, Christian Metzner, Laura Kriener, Melika Payvand

cs.LGcs.AIcs.NE

TLDR

NORACL uses neurogenesis to adaptively grow neural networks for continual learning, solving the oracle architecture problem and improving stability-plasticity.

Key contributions

  • Tackles stability-plasticity dilemma through adaptive neuronal growth.
  • Grows only when needed, monitoring representational and plasticity saturation.
  • Achieves competitive accuracy with fewer parameters than oracle-sized baselines.
  • Shows interpretable growth patterns based on task similarity.

Why it matters

NORACL solves the "oracle architecture problem" in continual learning by using neurogenesis to adaptively grow networks. This eliminates the need to pre-size models, enhancing stability, plasticity, and efficiency for lifelong learning.

Original Abstract

In a continual learning setting, we require a model to be plastic enough to learn a new task and stable enough to not disturb previously learned capabilities. We argue that this dilemma has an architectural root. A finite network has limited representational and plastic resources, yet the required capacity depends on properties of the future task stream that are unknown: how many tasks will be encountered, and how much they overlap in feature space. Regularization-based methods preserve past knowledge within fixed-capacity architectures and therefore implicitly rely on an oracle architecture sized for this unknown future. When tasks are only weakly related, fixed architectures progressively run out of plastic resources; when tasks are few or strongly overlapping, models are often over-provisioned. Inspired by neurogenesis in biology, we propose NORACL to address the stability-plasticity dilemma by tackling the oracle architecture problem through neuronal growth. Starting from a compact network, NORACL grows only when needed by monitoring two complementary signals for representational and plasticity saturation. We evaluate NORACL against oracle-sized static baselines across varying task counts and geometries. Across all settings, NORACL achieves final average accuracies that are better than or on par with oracle-provisioned static baselines while using fewer parameters. Additionally, NORACL yields architectures with interpretable growth, i.e. dissimilar tasks predominantly expand feature-extraction layers, whereas tasks which rely on common features shift growth toward later feature-combination layers. Our analysis further explains why fixed-capacity networks lose plasticity as tasks accumulate, whereas NORACL creates fresh capacity for new tasks through growth. Together, these results show that adaptive neurogenesis pushes the stability-plasticity Pareto frontier of continual learning.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.