Feature propagation as self-supervision signals on graphs

Self-supervised learning is gaining considerable attention as a solution to avoid the requirement of extensive annotations in representation learning on graphs. Current algorithms are based on contrastive learning, which is computationally and memory expensive, and the assumption of invariance under...

Full description

Saved in:
Bibliographic Details
Published inKnowledge-based systems Vol. 289; p. 111512
Main Authors Pina, Oscar, Vilaplana, Verónica
Format Journal Article
LanguageEnglish
Published Elsevier B.V 08.04.2024
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Self-supervised learning is gaining considerable attention as a solution to avoid the requirement of extensive annotations in representation learning on graphs. Current algorithms are based on contrastive learning, which is computationally and memory expensive, and the assumption of invariance under certain graph augmentations. However, graph transformations such as edge sampling may modify the semantics of the data, potentially leading to inaccurate invariance assumptions. To address these limitations, we introduce Regularized Graph Infomax (RGI), a simple yet effective framework for node level self-supervised learning that trains a graph neural network encoder by maximizing the mutual information between the output node embeddings and their propagation through the graph, which encode the nodes’ local and global context, respectively. RGI generates self-supervision signals through feature propagation rather than relying on graph data augmentations. Furthermore, the method is non-contrastive and does not depend on a two branch architecture. We run RGI on both transductive and inductive settings with popular graph benchmarks and show that it can achieve state-of-the-art performance regardless of its simplicity. •We present Regularized Graph Infomax (RGI) for efficient graph self-supervised learning.•RGI generates self-supervision signals with feature propagation.•The method does not use graph data augmentations and is non-contrastive.•State of the art performance is achieved in both transductive and inductive tasks.
ISSN:0950-7051
1872-7409
DOI:10.1016/j.knosys.2024.111512