MPCS: Neuroplastic Continual Learning via Multi-Component Plasticity and Topology-Aware EWC
TLDR
MPCS is a neuroplastic continual learning system integrating 11 mechanisms, achieving high efficiency and demonstrating critical component insights.
Key contributions
- Introduces MPCS, a neuroplastic system integrating 11 mechanisms for continual learning.
- Achieves 94.2 Normalized Efficiency Score on MEP-BENCH, a multi-domain benchmark.
- Identifies Fourier encoding as the single most critical component for performance.
- Finds that removing EWC entirely yields highest performance in high task-similarity.
Why it matters
This paper introduces MPCS, a highly efficient neuroplastic system for continual learning that balances stability and plasticity. It provides critical insights into architectural design, highlighting Fourier encoding's importance and EWC's nuanced role. The Pareto analysis also offers a practical guide for model compression.
Original Abstract
Continual learning systems face a fundamental tension between plasticity -- acquiring new knowledge -- and stability -- retaining prior knowledge. We introduce MPCS (Multi-Plasticity Continual System), a neuroplastic architecture that integrates eleven complementary mechanisms: task-driven neurogenesis, Fourier-encoded inputs, EWC regularization, meta-replay, mixed consolidation, hybrid gating, synapse pruning/regeneration, Hebbian updates, task similarity routing, adaptive growth control, and continuous neuron importance tracking. We evaluate MPCS on MEP-BENCH, a multi-track benchmark spanning 31 tasks across regression, classification, logic, and mixed domains, using a three-dimensional Pareto criterion over task performance (Perf), representation diversity (RD), and gradient conflict rate (GCR). Across 15 ablation configurations (3 seeds x 4 tracks x 2000 epochs), MPCS achieves a Normalized Efficiency Score of 94.2, placing it on the Pareto frontier among 9 of 14 gate-passing systems. Key findings: (i) Fourier encoding is the single most critical component (removal drops Perf by 30.7 pp and fails the MEP gate on 14% of tasks); (ii) global EWC degrades performance (NES = -4.2); topology-local EWC reduces this penalty (NES 90.5->91.8) but does not eliminate it; removing EWC entirely yields MPCS_EFFICIENT, the highest-Perf system -- establishing a monotone relationship in the high task-similarity regime (s_bar ~= 0.95): global EWC < topology EWC < no EWC; (iii) the Pareto status assessment is predictive: removing the two Pareto-dominated components (EWC + Hebbian) jointly yields MPCS_EFFICIENT, which improves Perf by 0.6 pp at 4.7x lower compute cost (127 vs. 602 min), validating the Pareto frontier as an actionable model-compression guide.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.