Sorry, you need to enable JavaScript to visit this website.

Noise-Disentangled Graph Contrastive Learning via Low-Rank and Sparse Subspace Decomposition

DOI:
10.60864/9rfd-b084
Citation Author(s):
Jiawei Sheng, Shicheng Wang, Tingwen Liu
Submitted by:
Gehang Zhang
Last updated:
6 June 2024 - 10:27am
Document Type:
Presentation Slides
Document Year:
2024
Event:
Presenters:
Gehang Zhang
Paper Code:
MLSP-L18
 

Graph contrastive learning aims to learn a representative model by maximizing the agreement between different views of the same graph. Existing studies usually allow multifarious noise in data augmentation, and suffer from trivial and inconsistent generation of graph views. Moreover, they mostly impose contrastive constraints on pairwise representations, limiting the structural correlations among multiple nodes. Both problems may hinder graph contrastive learning, leading to suboptimal node representations. To this end, we propose a novel graph contrastive learning framework, namely GCL-LS, via low-rank and sparse subspace decomposition. In particular, it decomposes node representations into low-rank and sparse components, preserving structural correlations and compressed features in the low-rank and sparse subspace, respectively. By contrasting the representations in the subspaces, it naturally disentangles low-quality noise in data augmentation, and captures structural correlations and substantial features of nodes in contrastive learning. Experimental results show that our method significantly improves downstream node classification accuracy, and further analysis demonstrates the effectiveness of the subspace decomposition in graph contrastive learning.

up
0 users have voted: