minAction.net: Energy-First Neural Architecture Design -- From Biological Principles to Systematic Validation
TLDR
minAction.net introduces energy-first neural architecture design, showing significant energy reduction and training efficiency gains with minimal accuracy loss.
Key contributions
- Optimal neural architecture is task-dependent, rejecting the universal best architecture assumption.
- Energy regularization reduces internal activation energy ~1000x with <0.5% accuracy change.
- Energy-first designs yield 5-33% training efficiency gains over conventional baselines.
Why it matters
This paper addresses the critical gap of energy efficiency in ML, showing that energy can be drastically reduced with minimal accuracy loss. It paves the way for more sustainable, biologically inspired AI systems by integrating energy constraints into design.
Original Abstract
Modern machine learning optimizes for accuracy without explicit treatment of internal computational cost, even though physical and biological systems operate under intrinsic energy constraints. We evaluate energy-aware learning across 2,203 experiments spanning vision, text, neuromorphic, and physiological datasets with 10 seeds per configuration and factorial statistical analysis. Three findings emerge. First, architecture alone explains negligible variance in accuracy (partial eta^2 = 0.001), while the architecture x dataset interaction is large (partial eta^2 = 0.44, p < 0.001), demonstrating that optimal architecture depends critically on task modality and rejecting the assumption of a universal best architecture. Second, a controlled lambda-sweep across lambda in {0, 1e-5, 1e-4, 1e-3, 1e-2} validates a single-parameter energy-regularized objective L = L_CE + lambda * E(theta, x): across this range, internal activation energy decreases by approximately three orders of magnitude relative to the unregularized lambda=0 baseline, with negligible accuracy change (<0.5 percentage points) on both MNIST and Fashion-MNIST. Third, energy-first architectures inspired by an action-principle framework yield 5-33% within-modality training-efficiency gains over conventional baselines. These results emerge from a research program that interprets learning through a structural correspondence between the action functional in classical mechanics, free energy in statistical physics, and KL-regularized objectives in variational inference. We frame this correspondence as a design hypothesis, not a derivation.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.