Interest-Aware Graph Contrastive Learning for Recommendation with Diffusion-based Augmentation

Mengyuan Jing, Yanmin Zhu, Zhaobo Wang, Jiadi Yu, Feilong Tang

Research output: Contribution to journalArticlepeer-review

Abstract

Graph Contrastive Learning (GCL) has recently garnered significant attention for enhancing recommender systems. Most existing GCL-based methods perturb the raw data graph to generate views, performing contrastive learning across these views to learn generalizable representations. However, most of these methods rely on data- or model-based augmentation techniques that may disrupt interest consistency. In this paper, we propose a novel interest-aware augmentation approach based on diffusion models to address this issue. Specifically, we leverage a conditional diffusion model to generate interest-consistent views by conditioning on node interaction information, ensuring that the generated views align with the interests of the nodes. Based on this augmentation method, we introduce DiffCL, a graph contrastive learning framework for recommendation. Furthermore, we propose an easy-to-hard generation strategy. By progressively adjusting the starting point of the reverse denoising process, this strategy further enhances effective contrastive learning. We evaluate DiffCL on three public real-world datasets, and results indicate that our method outperforms state-of-the-art techniques, demonstrating its effectiveness.

Original languageEnglish
JournalIEEE Transactions on Knowledge and Data Engineering
DOIs
StateAccepted/In press - 2025
Externally publishedYes

Keywords

  • contrastive learning
  • diffusion model
  • Graph-based Recommendation

Fingerprint

Dive into the research topics of 'Interest-Aware Graph Contrastive Learning for Recommendation with Diffusion-based Augmentation'. Together they form a unique fingerprint.

Cite this