DecompKAN: Decomposed Patch-KAN for Long-Term Time Series Forecasting
TLDR
DecompKAN is a lightweight, attention-free architecture combining decomposition, patching, and KANs for accurate and transparent long-term time series forecasting.
Key contributions
- DecompKAN: a lightweight, attention-free architecture for long-term time series forecasting.
- Integrates trend-residual decomposition, channel-wise patching, and learned instance normalization.
- Utilizes B-spline KANs for inspectable 1D scalar functions, enhancing model transparency.
- Achieves competitive MSE on benchmarks, excelling in smooth and physiological time series data.
Why it matters
DecompKAN offers a novel approach to long-term time series forecasting, balancing high accuracy with interpretability. Its unique architecture provides competitive results while allowing direct visualization of learned nonlinearities. This makes it valuable for scientific domains requiring both performance and transparency.
Original Abstract
Accurate time series forecasting in scientific domains such as climate modeling, physiological monitoring, and energy systems benefits from both competitive predictions and model transparency. This work proposes DecompKAN, a lightweight attention-free architecture that combines trend-residual decomposition, channel-wise patching, learned instance normalization, and B-spline Kolmogorov-Arnold Network (KAN) edge functions. Each KAN edge learns an explicit, inspectable 1D scalar function over learned patch-embedding coordinates that can be directly visualized. On standard benchmarks, DecompKAN achieves best or tied-best MSE on 15 of 32 dataset-horizon combinations among selected published baselines, and achieves best or tied-best MSE on 20 of 36 comparisons under a controlled same-recipe evaluation across 9 datasets including the physiological PPG-DaLiA benchmark. The architecture shows particular strength on datasets with smooth temporal dynamics (Solar -17%, ECL -10% vs. iTransformer, Weather) and physiological time series. Visualization of learned edge functions reveals qualitatively different latent nonlinearities across domains. Ablation analysis shows that the architectural pipeline (decomposition, patching, normalization) drives performance more than the choice of nonlinear layer, while the KAN formulation enables inspection of learned latent transformations.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.