ArXiv TLDR

ParaRNN: An Interpretable and Parallelizable Recurrent Neural Network for Time-Dependent Data

🐦 Tweet
2605.02692

Yuxi Cai, Lan Li, Feiqing Huang, Guodong Li

stat.MLcs.LG

TLDR

ParaRNN is a novel recurrent neural network designed for time-dependent data, offering enhanced interpretability and efficient parallelized training.

Key contributions

  • Introduces ParaRNN, a novel recurrent neural network composed of multiple small recurrent units.
  • Provides interpretability through an additive representation that decouples recurrent dynamics.
  • Enables efficient parallelization for faster training on time-dependent data.
  • Demonstrates comparable performance to vanilla RNNs with improved efficiency and interpretability.

Why it matters

This paper addresses key limitations of traditional RNNs: their lack of interpretability and slow training. ParaRNN offers a more transparent and efficient model for time-dependent data, bridging machine learning flexibility with statistical rigor. This advances the practical application of RNNs in complex statistical modeling.

Original Abstract

The proliferation of large-scale and structurally complex data has spurred the integration of machine learning methods into statistical modeling. Recurrent neural networks (RNNs), a foundational class of models for time-dependent data, can be viewed as nonlinear extensions of classical autoregressive moving average models. Despite their flexibility and empirical success in machine learning, RNNs often suffer from limited interpretability and slow training, which hinders their use in statistics. This paper proposes the Parallelized RNN (ParaRNN), a novel model composed of multiple small recurrent units. ParaRNN admits an additive representation that decouples recurrent dynamics into interpretable components, whose behavior can be characterized through recurrence features. This interpretability enables its applications in nonparametric regression for time-dependent data, while the design also allows efficient parallelization. The approximation capacity and non-asymptotic prediction error bounds in a nonparametric regression setting are established for ParaRNN. Empirical results on three sequential modeling tasks further demonstrate that ParaRNN achieves performance comparable to vanilla RNNs while offering improved interpretability and efficiency.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.