A Composite Activation Function for Learning Stable Binary Representations
Seokhun Park, Choeun Kim, Kwanho Lee, Sehyun Park, Insung Kong + 1 more
TLDR
Introduces HTAF, a smooth composite activation function enabling stable gradient-based training of neural networks with binary representations.
Key contributions
- Proposes Heavy Tailed Activation Function (HTAF) for stable training of binary neural networks.
- HTAF is a smooth composite function approximating Heaviside, enabling gradient-based optimization.
- Theoretically maintains large gradient mass around zero and slower decay in tail regions.
- Introduces Implicit Concept Bottleneck Models (ICBMs) for interpretable discrete image features.
Why it matters
Training neural networks with binary activations is challenging due to non-differentiability. This paper offers a novel, smooth activation function (HTAF) that resolves this, enabling stable gradient-based optimization. It also introduces ICBMs, providing a path towards more interpretable models with discrete features, advancing efficient and understandable AI.
Original Abstract
Activation functions play a central role in neural networks by shaping internal representations. Recently, learning binary activation representations has attracted significant attention due to their advantages in computational and memory efficiency, as well as interpretability. However, training neural networks with Heaviside activations remains challenging, as their non-differentiability obstructs standard gradient-based optimization. In this paper, we propose Heavy Tailed Activation Function (HTAF), a smooth approximation to the Heaviside function that enables stable training with gradient-based optimization. We construct HTAF as a sigmoid hyperbolic tangent composite function and theoretically show that it maintains a large gradient mass around zero inputs while exhibiting slower gradient decay in the tail regions. We show that Spiking Neural Networks, Binary Neural Networks and Deep Heaviside neural Networks can be trained stably using HTAF with gradient-based optimization. Finally, we introduce Implicit Concept Bottleneck Models (ICBMs), an interpretable image model that leverages HTAF to induce discrete feature representations. Extensive experiments across various architectures and image datasets demonstrate that ICBM enables stable discretization while achieving prediction performance comparable to or better than standard models.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.