Neural networks that learn from graph-structured data by aggregating information across connected nodes.
Graph Neural Networks (GNNs) are a class of deep learning models designed to operate directly on graph-structured data, where information is encoded not just in individual data points but in the relationships between them. Unlike conventional neural networks that expect inputs arranged in grids or sequences, GNNs work with nodes (entities) and edges (relationships), making them naturally suited to domains where connectivity is as meaningful as content — such as social networks, molecular structures, knowledge graphs, and transportation systems.
The core mechanism of a GNN is iterative neighborhood aggregation, sometimes called message passing. At each layer, every node collects feature information from its immediate neighbors, combines it with its own features using learned transformations, and updates its internal representation. Stacking multiple layers allows information to propagate across increasingly distant parts of the graph, enabling the model to capture both local structural patterns and broader topological context. Variants such as Graph Convolutional Networks (GCNs), Graph Attention Networks (GATs), and GraphSAGE differ primarily in how they weight and aggregate neighbor information — for instance, GATs use attention mechanisms to assign different importance to different neighbors.
GNNs gained significant traction following Thomas Kipf and Max Welling's 2016 paper on Graph Convolutional Networks, which offered a spectral-domain justification for neighborhood aggregation and demonstrated strong performance on semi-supervised node classification tasks. This work helped unify earlier, more fragmented approaches and sparked a wave of research that extended GNNs to edge prediction, graph classification, and dynamic graphs. Theoretical work on the expressive power of GNNs — particularly connections to the Weisfeiler-Leman graph isomorphism test — has since clarified both their capabilities and fundamental limitations.
In practice, GNNs have delivered meaningful results across a wide range of applications: predicting molecular properties in drug discovery, detecting fraud in financial transaction networks, powering recommendation engines at scale, and improving traffic forecasting. Their ability to model relational inductive biases directly makes them a powerful complement to CNNs and transformers in the broader deep learning toolkit, particularly as real-world datasets increasingly take the form of interconnected, heterogeneous graphs.