The Impact of Dimensionality on the Stability of Node Embeddings
Tobias Schumacher, Simon Reichelt, Markus Strohmaier
TLDR
This paper investigates how embedding dimensionality affects the stability and performance of five popular node embedding methods.
Key contributions
- Systematically evaluates 5 node embedding methods across varying dimensions.
- Reveals embedding stability significantly changes with dimensionality.
- Notes different stability trends for methods; some stabilize with higher dimensions.
- Highlights that maximum stability does not guarantee optimal task performance.
Why it matters
This work provides crucial insights into the trade-offs between stability, performance, and computational effectiveness in graph representation learning. It emphasizes the need for careful selection of embedding dimensions, as stability and performance are not always aligned.
Original Abstract
Previous work has established that neural network-based node embeddings return different outcomes when trained with identical parameters on the same dataset, just from using different training seeds. Yet, it has not been thoroughly analyzed how key hyperparameters such as embedding dimension could impact this instability. In this work, we investigate how varying the dimensionality of node embeddings influences both their stability and downstream performance. We systematically evaluate five widely used methods -- ASNE, DGI, GraphSAGE, node2vec, and VERSE -- across multiple datasets and embedding dimensions. We assess stability from both a representational perspective and a functional perspective, alongside performance evaluation. Our results show that embedding stability varies significantly with dimensionality, but we observe different patterns across the methods we consider: while some approaches, such as node2vec and ASNE, tend to become more stable with higher dimensionality, other methods do not exhibit the same trend. Moreover, we find that maximum stability does not necessarily align with optimal task performance. These findings highlight the importance of carefully selecting embedding dimension, and provide new insights into the trade-offs between stability, performance, and computational effectiveness in graph representation learning.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.