Scientific journal paper Q1
Quantum-enhanced learning: Leveraging von Neumann entropy for enhanced graph neural network performance
Muhammad Awais (Awais, M.); Octavian Postolache (Postolache, O. A.); Sancho Moura Oliveira (Oliveira, S. M.);
Journal Title
Neural Networks
Year (definitive publication)
2026
Language
English
Country
United States of America
More Information
Web of Science®

Times Cited: 0

(Last checked: 2026-04-26 22:02)

View record in Web of Science®

Scopus

Times Cited: 0

(Last checked: 2026-04-19 18:55)

View record in Scopus

Google Scholar

Times Cited: 0

(Last checked: 2026-04-26 17:40)

View record in Google Scholar

This publication is not indexed in Overton

Abstract
Graph Neural Networks (GNNs) have established themselves as powerful tools for learning from graph-structured data. However, their reliance on local message-passing mechanisms leads to over-squashing—the compression of exponentially growing neighborhood information into fixed-size vectors—which severely limits long-range dependency modeling. We introduce the Quantum-Inspired Graph Neural Network (QGNN) with a novel Quantum Entanglement Loss (QEL) function that addresses this challenge through a fundamentally different mechanism than existing approaches. Unlike spectral regularization (which enforces smoothness) or maximum entropy methods (which encourage representation diversity), QEL minimizes the von Neumann entropy of the node embedding correlation matrix, thereby concentrating eigenvalues in dominant eigenmodes that preserve global structural patterns. This entropy minimization creates direct information pathways between distant but functionally related nodes, effectively bypassing multi-hop bottlenecks. We evaluate QGNN on both standard benchmarks (Cora, Citeseer, PPI, Electronic Circuits) and the Long Range Graph Benchmark (LRGB) suite, which features graphs with average diameters up to 56.99 (Peptides). On LRGB datasets, QGNN achieves substantial improvements: 37.6% relative MAE reduction on Peptides-struct compared to GCN, 4.0% improvement over Graph Transformers (GraphGPS), and notably, 97% better performance than GCN on node pairs separated by 7+ hops. Despite these gains, QGNN requires only 20–30% additional computational overhead compared to standard GCN, while being 5–6 ×  faster than Graph Transformer approaches. Our results establish entropy-based regularization as a principled and efficient approach for long-range dependency modeling in graphs.
Acknowledgements
--
Keywords
Graph neural networks,von Neumann entropy Quantum information theory,Over-squashing,Over-smoothing
  • Computer and Information Sciences - Natural Sciences
  • Basic Medicine - Medical and Health Sciences
Funding Records
Funding Reference Funding Entity
UID/50008: Instituto de Telecomunicações Fundação para a Ciência e a Tecnologia
101,083,896 DIGITAL-2021-SKILLS-01 Comissão Europeia