Spatiotemporal Convolutions on EEG signal -- A Representation Learning Perspective on Efficient and Explainable EEG Classification with Convolutional Neural Nets
Laurits Dixen, Stefan Heinrich, Paolo Burelli
TLDR
2D spatiotemporal convolutions on EEG signals significantly reduce training time for high-dimensional classification tasks while maintaining performance.
Key contributions
- Compared 1D and 2D spatiotemporal convolutions for EEG signal classification.
- Found 2D convolutions significantly reduce training time in high-dimensional tasks.
- 2D convolutions maintain classification performance comparable to 1D methods.
- Revealed 1D and 2D models create vastly different internal representational geometries.
Why it matters
This paper offers a more efficient architectural choice for EEG classification, significantly reducing training time. It highlights the importance of understanding internal representations, not just performance, for designing CNNs, potentially accelerating EEG-based BCI development.
Original Abstract
Classification of EEG signals using shallow Convolutional Neural Networks (CNNs) is a prevalent and successful approach across a variety of fields. Most of these models use independent one-dimensional (1D) convolutional layers along the spatial and temporal dimensions, which are concatenated without a non-linear activation layer between. In this paper, we investigate an alternative encoding that operates a bi-dimensional (2D) spatiotemporal convolution. While 2D convolutions are numerically identical to two concatenated 1D convolutions along the two dimensions, the impact on learning is still uncertain. We test 1D and 2D CNNs and a CNN+transformer hybrid model in a low-dimensional (3-channel) and a high-dimensional (22-channel) BCI motor imagery classification task. We observe that 2D convolutions significantly reduce training time in high-dimensional tasks while maintaining performance. We investigate the root of this improvement and find no difference in spectral feature importance. However, a clear pattern emerges in representational similarity across models: 1D and 2D models yield vastly different representational geometries. Overall, we suggest an improved model with a 2D convolutional layer for faster training and inference. We also highlight the importance of architecturally-driven encoding when processing complex multivariate signals, as reflected in internal representations rather than purely in performance metrics.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.