This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Denoised Self-Augmented Learning for Social Recommendation

Tianle Wang, Lianghao Xia, Chao Huang
University of Hong Kong, Hong Kong
louiswong.cs@connect.hku.hk, aka_xia@foxmail.com
chaohuang75@gmail.com
Chao Huang is the Corresponding Author
Abstract

Social recommendation is gaining increasing attention in various online applications, including e-commerce and online streaming, where social information is leveraged to improve user-item interaction modeling. Recently, Self-Supervised Learning (SSL) has proven to be remarkably effective in addressing data sparsity through augmented learning tasks. Inspired by this, researchers have attempted to incorporate SSL into social recommendation by supplementing the primary supervised task with social-aware self-supervised signals. However, social information can be unavoidably noisy in characterizing user preferences due to the ubiquitous presence of interest-irrelevant social connections, such as colleagues or classmates who do not share many common interests. To address this challenge, we propose a novel social recommender called the Denoised Self-Augmented Learning paradigm (DSL). Our model not only preserves helpful social relations to enhance user-item interaction modeling but also enables personalized cross-view knowledge transfer through adaptive semantic alignment in embedding space. Our experimental results on various recommendation benchmarks confirm the superiority of our DSL over state-of-the-art methods. We release our model implementation at: https://github.com/HKUDS/DSL.

1 Introduction

Social recommendation is a widely-used technique to improve the quality of recommender systems by incorporating social information into user preference learning Yu et al. (2021a). To accomplish this, various neural network techniques have been developed to encode social-aware user preferences for recommendation. Currently, the most advanced social recommendation methods are built using Graph Neural Networks (GNNs) for recursive message passing, which enables the capture of high-order correlations Fan et al. (2019); Wu et al. (2019); Song et al. (2019). In these architectures, user representations are refined by integrating information from both social and interaction neighbors.

While supervised GNN-enhanced models have achieved remarkable performance in social recommendation, they require a large amount of supervised labels to generate accurate user representations. In practical social recommendation scenarios, however, user-item interaction data is often very sparse Wei et al. (2022a); Chen et al. (2023). This label sparsity severely limits the representation power of deep social recommenders and hinders their ability to reach their full potential. Recently, Self-Supervised Learning (SSL) has gained success due to its ability to avoid heavy reliance on observed label data in various domains, e.g., computer vision He et al. (2020a), natural language processing Liu and Liu (2021), and graph representation learning Zhu et al. (2021).

Motivated by the limitations of supervised GNN-enhanced models, recent attempts have adopted the self-supervised learning framework Yu et al. (2021a). These approach introduce an auxiliary learning task to supplement the supervised main task for data augmentation. For example, MHCN Yu et al. (2021b) uses a hypergraph-enhanced self-supervised learning framework to improve global relation learning in social recommender systems. Additionally, SMIN Long et al. (2021) constructs metapath-guided node connections to explore the isomorphic transformation property of graph topology with augmented self-supervision signals.

Despite the decent performance of self-supervised learning, we argue that the SSL-based augmentation is severely hindered by noisy social relations when enhancing the representation learning of complex user preferences. While observed user-user social ties have the potential to capture social influence on user-item interaction behaviors, the model’s performance can significantly degrade when trained on social-aware collaborative graphs with noisy social information. For instance, people may establish social connections with colleagues, classmates, or family members, but they may not share many common interests with each other Liu et al. (2019); Epasto and Perozzi (2019). Therefore, these noisy social influences may not align with user preferences in real-life recommendation scenarios. In most existing solutions, information aggregated from noisy social neighbors may mislead graph message passing and self-supervised learning, resulting in sub-optimal recommendation performance.

To address the limitations mentioned earlier, we propose the Denoised Self-Augmented Learning (DSL) paradigm for social recommender systems. Our approach leverages social information to better characterize user preferences with noise-resistant self-supervised learning, aimed at pursuing cross-view alignment. Firstly, we develop a dual-view graph neural network to encode latent representations over both user social and interaction graphs. Then, to mitigate the bias of social relations for recommendation, we design a denoising module to enhance the integrated social-aware self-supervised learning task. This module identifies unreliable user-wise connections with respect to their interaction patterns. Our DSL is aware of the interaction commonality between users and can automate the social effect denoising process with adaptive user representation alignment. By doing so, the social-aware uniformity is well preserved in the learned user embeddings by alleviating the impact of noisy social information in our recommender.

Key contributions of this work are summarized as follows:

  • In this work, we investigate denoised self-augmented learning for social recommendation, effectively reducing the impact of noisy social relations on the representation of socially-aware collaborative signals.

  • We propose DSL, which enables denoised cross-view alignment between the encoded embeddings from social and interaction views. The denoising module assigns reliability weights to useful social relations to encode user preference, endowing our DSL with the capability of generating adaptive self-supervised signals.

  • We instantiate DSL for social recommendation on three real-world datasets. Experimental results show that our method provides more accurate recommendations and superior performance in dealing with noisy and sparse data, compared with various state-of-the-art solutions.

2 Preliminaries and Related Work

2.1 Social-aware Recommendation

We denote the sets of users and items as 𝒰={u1,,uI}\mathcal{U}=\{u_{1},...,u_{I}\} and 𝒱={v1,,vJ}\mathcal{V}=\{v_{1},...,v_{J}\}, respectively, where II and JJ represent the number of users and items. The user-item interaction data is represented by an interaction graph 𝒢r={𝒰,𝒱,r}\mathcal{G}_{r}=\{\mathcal{U},\mathcal{V},\mathcal{E}_{r}\}, where user-item connection edges are generated when user uiu_{i} interacts with item vjv_{j}. To incorporate social context into the recommender system, we define a user social graph 𝒢s={𝒰,s}\mathcal{G}_{s}=\{\mathcal{U},\mathcal{E}_{s}\} to contain user-wise social connections. Social recommender systems aim to learn a model from both the user-item interaction graph 𝒢r={𝒰,𝒱,r}\mathcal{G}_{r}=\{\mathcal{U},\mathcal{V},\mathcal{E}_{r}\} and the user-user social graph 𝒢s={𝒰,s}\mathcal{G}_{s}=\{\mathcal{U},\mathcal{E}_{s}\}, encoding user interests to make accurate recommendations.

In recent years, several neural network techniques have been proposed to solve the social recommendation problem. For instance, attention mechanisms have been used to differentiate social influence among users with learned attentional weights, as seen in EATNN Chen et al. (2019b) and SAMN Chen et al. (2019a). Many GNN-based social recommender systems have been developed to jointly model user-user and user-item graphs via message passing, leveraging the effectiveness of high-order relation encoding with graph neural networks Wang et al. (2019). Examples include GraphRec Fan et al. (2019), DiffNet Wu et al. (2019), and FeSoG Liu et al. (2022). Some recent attempts have leveraged self-supervised learning to enhance social recommendation with auxiliary self-supervision signals, such as MHCN Yu et al. (2021b) and SMIN Long et al. (2021). However, their representation learning abilities are limited by social relation noise, which leads to biased models.

2.2 Self-Supervised Recommender Systems

Self-supervised learning has recently gained attention in various recommendation tasks. Supervised contrastive loss has been shown to benefit graph collaborative filtering with effective data augmentation Wu et al. (2021); Cai et al. (2023). For sequential recommender systems, self-supervised pre-training Zhou et al. (2020) and imitation Yuan et al. (2022) have been introduced to enhance sequence modeling. Researchers have also brought the benefits of self-supervised learning to multi-interest/multi-behavior recommender systems, as seen in Re4 Zhang et al. (2022) and CML Wei et al. (2022b). Our method advances this research by proposing a novel unbiased self-supervised learning paradigm to denoise social relation modeling in social recommender systems.

3 Methodology

Refer to caption
Figure 1: Overall framework of the proposed denoised self-augmented learning (DSL) model.

In this section, we present our DSL model with technical details. The model architecture is illustrated in Figure 1.

3.1 Dual-View Graph Neural Relation Learning

With the initialized id-corresponding embeddings, our DSL first employs a dual-view graph neural network to capture high-order collaborative relations for both user-item interactions and user-user social ties. Inspired by the effectiveness of lightweight GCN-enhanced collaborative filtering paradigms He et al. (2020b); Chen et al. (2020), DSL is configured with a simplified graph neural network, which is:

Er(l)=(r+I)Er(l1)\displaystyle\textbf{E}^{(l)}_{r}=(\mathcal{L}_{r}+\textbf{I})\cdot\textbf{E}^{(l-1)}_{r} (1)

The above equation shows the iterative information propagation scheme of our GCN over the user-item interaction graph. Here, Er(l),Er(l1)(I+J)×d\textbf{E}^{(l)}_{r},\textbf{E}_{r}^{(l-1)}\in\mathbb{R}^{(I+J)\times d} denote the embeddings of users and items after ll iterations of user-item relation modeling. Er(0)\textbf{E}^{(0)}_{r} is initialized by stacking the initial user embedding matrix Eu\textbf{E}_{u} and the item embedding matrix Ev\textbf{E}_{v}. I(I+J)×(I+J)\textbf{I}\in\mathbb{R}^{(I+J)\times(I+J)} denotes the identity matrix for enabling self-loop. r(I+J)×(I+J)\mathcal{L}_{r}\in\mathbb{R}^{(I+J)\times(I+J)} denotes the Laplacian matrix of the user-item interaction graph Wang et al. (2019).

r=Dr12ArDr12,Ar=[0RR0]\displaystyle\mathcal{L}_{r}=\textbf{D}_{r}^{-\frac{1}{2}}\textbf{A}_{r}\textbf{D}_{r}^{-\frac{1}{2}},\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\ \textbf{A}_{r}=\left[\begin{array}[]{cc}\textbf{0}&\textbf{R}\\ \textbf{R}^{\top}&\textbf{0}\end{array}\right] (4)

RI×J\textbf{R}\in\mathbb{R}^{I\times J} denotes the user-item interaction matrix, and 0 refers to all-zero matrices. The bidirectional adjacent matrix Ar\textbf{A}_{r} of the user-item interaction view is multiplied by its corresponding diagonal degree matrix Dr\textbf{D}_{r} for normalization.

To encode user-wise social relations in our recommender, we also apply the lightweight GCN to the user social graph 𝒢s\mathcal{G}_{s}. Specifically, our social view GNN takes the initial users’ id-corresponding embeddings as input by setting Es(0)=Eu\textbf{E}^{(0)}_{s}=\textbf{E}_{u}. The user embeddings are generated by cross-layer passing:

Es(l)=(s+I)Es(l1),s=Ds12SDs12\displaystyle\textbf{E}^{(l)}_{s}=(\mathcal{L}_{s}+\textbf{I})\cdot\textbf{E}_{s}^{(l-1)},\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\ \mathcal{L}_{s}=\textbf{D}_{s}^{-\frac{1}{2}}\textbf{S}\leavevmode\nobreak\ \textbf{D}_{s}^{-\frac{1}{2}} (5)

Here, SI×I\textbf{S}\in\mathbb{R}^{I\times I} encodes the user-wise social relations, and Ds,sI×I\textbf{D}_{s},\mathcal{L}_{s}\in\mathbb{R}^{I\times I} denote the corresponding diagonal degree matrix and the normalized Laplacian matrix for the social view. Es(l),Es(l1)I×d\textbf{E}_{s}^{(l)},\textbf{E}_{s}^{(l-1)}\in\mathbb{R}^{I\times d} are the users’ social embeddings in the ll-th and (l1)(l-1)-th graph neural iteration, respectively.

Embedding Aggregation.

To aggregate the embeddings encoded from different orders in 𝒢r\mathcal{G}_{r} and 𝒢s\mathcal{G}_{s}, DSL adopts mean-pooling operators for both the interaction and social views.

E¯r=l=0LEr(l),E¯s=l=0LEs(l)\displaystyle\bar{\textbf{E}}_{r}=\sum_{l=0}^{L}\textbf{E}_{r}^{(l)},\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\ \bar{\textbf{E}}_{s}=\sum_{l=0}^{L}\textbf{E}_{s}^{(l)} (6)

Here, LL is the maximum number of graph iterations. With our dual-view GNN, we encode view-specific relations for user interaction and social influence in our model.

3.2 Cross-View Denoised Self-Supervision

In our recommender, the learned user-item relations and user-wise dependencies are complementary to each other. To integrate both relational contextual signals, we design a cross-view denoised self-supervised learning paradigm that can alleviate the noisy effects of transferring social knowledge into user-item interaction modeling. In real-life scenarios, passively-built social relations, such as colleagues or classmates, may not bring much influence to user interaction preference due to their diverse shopping tastes. Blindly relying on such irrelevant social ties to infer users’ interests could damage the performance of social recommendation models. To address this issue, we filter out the noisy social influence between dissimilar users with respect to their interaction preference for unbiased self-supervision.

Adaptive Cross-View Alignment.

In our DSL, we incorporate the cross-view denoising task to supplement the main learning task with auxiliary self-supervision signals. The learned user interaction patterns guide the social relation denoising module to filter out misleading embedding propagation based on observed social connections. Specifically, the interaction similarity zi,iz_{i,i^{\prime}} between the user pair (ii, ii^{\prime}) is generated by zi,i=[e¯i(r);e¯i(r)]z_{i,i^{\prime}}=[\bar{\textbf{e}}^{(r)}_{i};\bar{\textbf{e}}^{(r)}_{i^{\prime}}], given the user embeddings (e¯i(r)\bar{\textbf{e}}^{(r)}_{i}, e¯i(r)\bar{\textbf{e}}^{(r)}_{i^{\prime}}) learned from our interaction GNN. Similarly, user social similarity z^i,i\hat{z}_{i,i^{\prime}} can be obtained by z^i,i=[e¯i(s);e¯i(s)]\hat{z}_{i,i^{\prime}}=[\bar{\textbf{e}}_{i}^{(s)};\bar{\textbf{e}}_{i^{\prime}}^{(s)}], based on user representations (e¯i(s)\bar{\textbf{e}}^{(s)}_{i}, e¯i(s)\bar{\textbf{e}}^{(s)}_{i^{\prime}}) encoded from our social GNN. To alleviate the semantic gap between interaction view and social view, we design a learnable similarity projection function to map interaction semantics into a latent embedding space for cross-view alignment, as follows:

zi,i=sigm(dσ(T[e¯i(r);e¯i(r)]+e¯i(r)+e¯i(r)+c))\displaystyle z_{i,i^{\prime}}=\text{sigm}(\textbf{d}^{\top}\cdot\sigma(\textbf{T}\cdot[\bar{\textbf{e}}^{(r)}_{i};\bar{\textbf{e}}^{(r)}_{i^{\prime}}]+\bar{\textbf{e}}^{(r)}_{i}+\bar{\textbf{e}}^{(r)}_{i^{\prime}}+\textbf{c})) (7)

where sigm()\text{sigm}(\cdot) and σ()\sigma(\cdot) denote the sigmoid and LeakyReLU activation functions, respectively. Our designed parameterized projection function consists of dd,Td×2d,cd\textbf{d}\in\mathbb{R}^{d},\textbf{T}\in\mathbb{R}^{d\times 2d},\textbf{c}\in\mathbb{R}^{d} as learnable parameters, enabling adaptive alignment between social and interaction views.

Denoised Self-Supervised Augmentation.

To incorporate denoised social influence to improve recommendation quality, we design a self-supervised learning task for cross-view alignment with augmented embedding regularization. Specifically, the cross-view alignment loss function is:

Lssl=(ui,ui)max(0,1zi,iz^i,i)\displaystyle L_{ssl}=\sum_{(u_{i},u_{i^{\prime}})}\max(0,1-z_{i,i^{\prime}}\hat{z}_{i,i^{\prime}}) (8)

The user pair (ui,uiu_{i},u_{i^{\prime}}) is individually sampled from user set 𝒰\mathcal{U}. With the above self-supervised learning objective, the integrated user relation prediction task will be guided based on the self-supervised signals for social influence denoising. Dy doing so, the noisy social connections between users with dissimilar preference, which contradicts with the target recommendation task will result in distinguishable user representations for recommendation enhancement.

3.3 Multi-Task Model Optimization

The learning process of our DSL involves multi-task training for model optimization. The augmented self-supervised learning tasks are integrated with the main recommendation optimized loss to model denoised social-aware user preferences. Given the encoded user and item embeddings, we predict user-item (y^ui,vj(r)\hat{y}_{u_{i},v_{j}}^{(r)}) and user-user (y^ui,ui(s)\hat{y}_{u_{i},u_{i^{\prime}}}^{(s)}) relations as:

y^ui,vj(r)=e¯i(r)e¯j(r);y^ui,ui(s)=e¯i(s)e¯i(s)\displaystyle\hat{y}_{u_{i},v_{j}}^{(r)}=\bar{\textbf{e}}_{i}^{(r)\top}\bar{\textbf{e}}_{j}^{(r)};\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\ \hat{y}_{u_{i},u_{i^{\prime}}}^{(s)}=\bar{\textbf{e}}_{i}^{(s)\top}\bar{\textbf{e}}_{i^{\prime}}^{(s)} (9)

where y^ui,vj(r)\hat{y}_{u_{i},v_{j}}^{(r)}\in\mathbb{R} represents the likelihood of user uiu_{i} interacting with item vjv_{j} from the interaction view, while y^ui,ui(s)\hat{y}_{u_{i},u_{i^{\prime}}}^{(s)} indicates the probability of uiu_{i} and uiu_{i^{\prime}} being socially connected. Given these definitions, we minimize the following BPR loss functions Rendle et al. (2009) for optimization:

Lrec=(ui,vj+,vj)lnsigm(y^ui,vj+(r)y^ui,vj(r))\displaystyle{L}_{rec}=\sum_{(u_{i},v_{j^{+}},v_{j^{-}})}-\ln\text{sigm}(\hat{y}_{u_{i},v_{j^{+}}}^{(r)}-\hat{y}_{u_{i},v_{j^{-}}}^{(r)})
Lsoc=(ui,ui+,ui)lnsigm(y^ui,ui+(s)y^ui,ui(s))\displaystyle{L}_{soc}=\sum_{(u_{i},u_{i^{+}},u_{i^{-}})}-\ln\text{sigm}(\hat{y}_{u_{i},u_{i^{+}}}^{(s)}-\hat{y}_{u_{i},u_{i^{-}}}^{(s)}) (10)

where vj+v_{j^{+}} and vjv_{j^{-}} denote the sampled positive and negative item for user uiu_{i}. ui+u_{i^{+}} and uiu_{i^{-}} are sampled from uiu_{i}’s socially-connected and unconnected users, respectively. By integrating self-supervised learning objectives with weight parameter λ1,λ2,λ3\lambda_{1},\lambda_{2},\lambda_{3}, the joint optimized loss is given as:

L=Lrec+λ1Lsoc+λ2Lssl+λ3(EuF2+EvF2)\displaystyle L=L_{rec}+\lambda_{1}L_{soc}+\lambda_{2}L_{ssl}+\lambda_{3}(\|\textbf{E}_{u}\|_{F}^{2}+\|\textbf{E}_{v}\|_{F}^{2}) (11)

3.4 In-Depth Analysis of DSL

In this section, our aim is to answer the question: How does our model enable adaptive and efficient self-supervised learning? We provide analysis to further understand our model.

Adaptive Self-Supervised Learning.

In most existing contrastive learning (CL) approaches, auxiliary SSL signals are generated to address the issue of sparse supervision labels. Following the mutual information maximization (Infomax) principle, these approaches maximize the agreement between positive samples while pushing negative pairs away in the embedding space, as shown below with derived gradients.

Lcle¯i=e¯i++uie¯iexpe¯ie¯iuiexpe¯ie¯i\displaystyle\frac{\partial L_{cl}}{\partial\bar{\textbf{e}}_{i}}=-\bar{\textbf{e}}_{i^{+}}+\sum_{u_{i^{-}}}\bar{\textbf{e}}_{i^{-}}\frac{\exp\bar{\textbf{e}}_{i}^{\top}\bar{\textbf{e}}_{i^{-}}}{\sum_{u_{i^{-}}}\exp\bar{\textbf{e}}_{i}^{\top}\bar{\textbf{e}}_{i^{-}}} (12)

The first term in the equation maximizes the similarity between positive pairs (e¯i\bar{\textbf{e}}_{i} and e¯i+\bar{\textbf{e}}_{i^{+}}) with the same strength. Negative pairs (e¯i\bar{\textbf{e}}_{i} and e¯i\bar{\textbf{e}}_{i^{-}}) with higher similarity are pushed away with greater strength, while negative pairs with lower similarity are pushed away with lesser strength. For simplicity, we have omitted the vector normalization and the temperature coefficient in the above-presented InfoNCE loss.

In comparison, our cross-view denoising SSL schema aims to maximize the similarity between sampled user pairs (e¯i,e¯i)(\bar{\textbf{e}}_{i},\bar{\textbf{e}}_{i^{\prime}}) adaptively, based on the labels zi,iz_{i,i^{\prime}}. The non-zero gradients of our denoising SSL over e¯i\bar{\textbf{e}}_{i} are shown below:

Lssle¯i=uizi,ie¯i\displaystyle\frac{\partial L_{ssl}}{\partial\bar{\textbf{e}}_{i}}=\sum_{u_{i^{\prime}}}-z_{i,i^{\prime}}\bar{\textbf{e}}_{i^{\prime}} (13)

The learnable zi,iz_{i,i^{\prime}} reflects the common preference between user uiu_{i} and uiu_{i^{\prime}}, which adaptively controls the strength of our self-supervised regularization. This enables us to filter out noisy signals in the observed social connections, and supercharge our SSL paradigm with adaptive data augmentation by transferring knowledge across different semantic views.

Efficient SSL.

Our DSL model adopts a lightweight graph convolutional network (GCN) as the graph relation encoder, with a complexity of 𝒪((|r|+|s|)×d)\mathcal{O}((|\mathcal{E}_{r}|+|\mathcal{E}_{s}|)\times d). The cross-view denoised self-supervised learning conducts pairwise node-wise relationships, which takes 𝒪(B×d)\mathcal{O}(B\times d) time complexity, where BB represents the batch size. In contrast, most existing vanilla InfoNCE-based contrastive learning methods calculate relations between a batch of nodes and all other nodes, resulting in an operation complexity of 𝒪(B×I×d)\mathcal{O}(B\times I\times d).

4 Evaluation

Data Ciao Epinions Yelp
# Users 6,672 11,111 161,305
# Items 98,875 190,774 114,852
# Interactions 198,181 247,591 1,118,645
Interaction Density 0.0300% 0.0117% 0.0060%
# Social Ties 109,503 203,989 2,142,242
Table 1: Statistical information of evaluated datasets.

We conduct extensive experiments to evaluate the effectiveness of our DSL by answering the following research questions: RQ1: Does DSL outperform state-of-the-art recommender systems? RQ2: How do different components affect the performance of DSL? RQ3: Is DSL robust enough to handle noisy and sparse data in social recommendation? RQ4: How efficient is DSL compared to alternative methods?

4.1 Experimental Settings

Dataset.

We conduct experiments on three benchmark datasets collected from the Ciao, Epinions, and Yelp online platforms, where social connections can be established among users in addition to their observed implicit feedback (e.g., rating, click) over different items. Table 1 lists the detailed statistical information of the experimented datasets.

Metrics.

We use Hit Ratio (HR)@N and Normalized Discounted Cumulative Gain (NDCG)@N as evaluation metrics, where NN is set to 10 by default. We adopt a leave-one-out strategy, following similar settings as in Long et al. (2021).

Dataset Metrics PMF TrustMF DiffNet DGRec EATNN NGCF+ MHCN KCGN SMIN DSL %Imp
Ciao HR 0.4223 0.4492 0.5544 0.4658 0.4255 0.5629 0.5950 0.5785 0.5852 0.6374 26.0
NDCG 0.2464 0.2520 0.3167 0.2401 0.2525 0.3429 0.3805 0.3552 0.3687 0.4065 37.2
Epinions HR 0.1686 0.1769 0.2182 0.2055 0.1576 0.2969 0.3507 0.3122 0.3159 0.3983 76.9
NDCG 0.0968 0.0842 0.1162 0.0908 0.0794 0.1582 0.1926 0.1721 0.1867 0.2290 96.2
Yelp HR 0.7554 0.7791 0.8031 0.7950 0.8031 0.8265 0.8571 0.8484 0.8478 0.8923 10.1
NDCG 0.5165 0.5424 0.5670 0.5593 0.5560 0.5854 0.6310 0.6028 0.5993 0.6599 15.5
Table 2: Recommendation performance of different methods. %Imp denotes relative improvements over all baselines on average.

4.2 Baseline Methods

We evaluate the performance of DSL by comparing it with 10 baselines from different research lines for comprehensive evaluation, including: i) MF-based recommendation approaches (i.e., PMF, TrustMF); ii) attentional social recommenders (i.e., EATNN); iii) GNN-enhanced social recommendation methods (i.e., DiffNet, DGRec, NGCF+); and iv) self-supervised social recommendation models (i.e., MHCN, KCGN, SMIN, DcRec). Details are provided as follows:

  • PMF Mnih and Salakhutdinov (2007): is a probabilistic approach that uses matrix factorization technique to factorize users and items into latent vectors for representations.

  • TrustMF Yang et al. (2016): This method incorporates trust relations between users into matrix factorization as social information to improve recommendation performance.

  • EATNN Chen et al. (2019b): It is an adaptive transfer learning model built upon attention mechanisms to aggregate information from both user interactions and social ties.

  • DiffNet Wu et al. (2019): This is a deep influence propagation architecture to recursively update users’ embeddings with social influence diffusion components.

  • DGRec Song et al. (2019): This social recommender leverages a graph attention network to jointly model the dynamic behavioral patterns of users and social influence.

  • NGCF+ Wang et al. (2019): This GNN-enhanced collaborative filtering approach performs message passing over a social-aware user-item relation graph.

  • MHCN Yu et al. (2021b): This model proposes a multi-channel hypergraph convolutional network to enhance social recommendation by considering high-order relations.

  • KCGN Huang et al. (2021): It improves social recommendation by integrating item inter-dependent knowledge with social influence through a multi-task learning framework.

  • SMIN Long et al. (2021): This model incorporates a metapath-guided heterogeneous graph learning task into social recommendation, utilizing self-supervised signals based on mutual information maximization.

Implementation Details.

We implement our DSL using PyTorch and optimize parameter inference with Adam. During training, we use a learning rate range of [5e4,1e3,5e3][5e^{-4},1e^{-3},5e^{-3}] and a decay ratio of 0.960.96 per epoch. The batch size is selected from [1024,2048,4096,8192][1024,2048,4096,8192] and the hidden dimensionality is tuned from [64,128,256,512][64,128,256,512]. We search for the optimal number of information propagation layers in our graph neural architecture from [1,2,3,4][1,2,3,4]. The regularization weights λ1\lambda_{1} and λ2\lambda_{2} are selected from [1e3,1e2,1e1,1e0,1e1][1e^{-3},1e^{-2},1e^{-1},1e^{0},1e^{1}] and [1e6,1e5,1e4,1e3][1e^{-6},1e^{-5},1e^{-4},1e^{-3}], respectively. The weight for weight-decay regularization λ3\lambda_{3} is tuned from [1e7,1e6,1e5,1e4][1e^{-7},1e^{-6},1e^{-5},1e^{-4}].

4.3 Overall Performance Comparison (RQ1)

Table 2 and Table 4 demonstrate that our DSL framework consistently outperforms all baselines on various datasets, providing evidence of its effectiveness. Based on our results, we make the following observations.

  • Our results demonstrate that DSL achieves encouraging improvements on datasets with diverse characteristics, such as varying interaction densities. Specifically, on the Ciao and Epinions datasets, DSL achieves an average improvement of 26.0% and 76.0% over baselines, respectively. This validates the importance of addressing noise issues in social information incorporation, which can effectively debias user representations and boost recommendation performance.

  • Methods incorporating self-supervised augmentation consistently outperform other baselines, highlighting the importance of exploring self-supervision signals from unlabeled data to alleviate sparsity issues in social recommendation. Our DSL outperforms other methods, suggesting that denoising social relation modeling in socially-aware recommender systems can benefit the design of more helpful self-supervision information. In MHCN and SMIN, mutual information is maximized to reach agreement between socially-connected users under a self-supervised learning framework. However, blindly generating augmented self-supervision labels from noisy social connections can align embeddings of connected users, diluting their true preferences. In contrast, our DSL can mitigate the effects of false positives for socially-dependent users.

  • GNN-enhanced social recommenders, e.g., DiffNet and DGRec, outperform vanilla attentive methods like EATNN, highlighting the effectiveness of modeling high-order connectivity in social-aware collaborative relationships. This observation aligns with the conclusion that incorporating high-hop information fusion is beneficial for embedding learning in CF signals. However, aggregating irrelevant social information via GNNs can lead to unwanted embedding propagation and weaken model representation ability.

Data Ciao Epinions Yelp
Metrics HR NDCG HR NDCG HR NDCG
DSL-d 0.615 0.399 0.354 0.207 0.887 0.658
DSL-s 0.594 0.374 0.327 0.169 0.839 0.621
DSL-c 0.603 0.388 0.336 0.199 0.889 0.662
DSL 0.637 0.406 0.398 0.229 0.892 0.659
Table 3: Component-wise ablation study of DSL.
Metrics PMF TrustMF DiffNet DGRec EATNN NGCF+ MHCN KCGN SMIN DSL %Imp
HR@5 0.3032 0.3133 0.3963 0.3035 0.3259 0.4263 0.4762 0.4361 0.4565 0.5007 35.2
NDCG@5 0.2071 0.2073 0.2650 0.1872 0.2342 0.3020 0.3479 0.3094 0.3298 0.3626 43.0
HR@10 0.4223 0.4492 0.5544 0.4658 0.4255 0.5629 0.5950 0.5785 0.5852 0.6374 26.0
NDCG@10 0.2464 0.2520 0.3167 0.2401 0.2525 0.3429 0.3805 0.3552 0.3687 0.4065 37.2
HR@20 0.5565 0.6133 0.6973 0.6193 0.5309 0.7032 0.7418 0.7191 0.7000 0.7683 18.6
NDCG@20 0.2799 0.3020 0.3514 0.2746 0.2838 0.3675 0.4241 0.3844 0.3780 0.4353 31.6
Table 4: Ranking performance on Ciao dataset with varying Top-N values in terms of HR@N and NDCG@N

4.4 Impact Study of Different Components (RQ2)

In this section, we examine the effects of component-wise ablation and how hyperparameters influence performance.

Model Ablation Study.

To investigate the essential role of our denoised self-supervised learning paradigm in improving performance, we perform an ablation study of key model components. Specifically, we compare our DSL with the following ablated variants: (1) “DSL-d”: disabling cross-view denoised self-supervised learning for mitigating the noisy effects of social-aware collaborative filtering, (2) “DSL-s”: removing social-aware self-supervised augmentation and directly incorporating social user embeddings into user-item interaction prediction, and (3) “DSL-c”: replacing denoised cross-view alignment with contrastive learning to reach agreement for integrating social and interaction views. Results are reported in Table 3. From our results, we observe that our DSL outperforms other variants in most evaluation cases. Based on this, we draw the following key conclusions:

  • Comparing DSL with “DSL-d”, the significant performance improvement suggests that social-aware collaborative relation learning is affected by the noise problem when directly incorporating social information.

  • The recommendation performance further drops in variant “DSL-s” without the auxiliary learning task of social-aware relation prediction, indicating that incorporating self-supervised signals from user-wise social influence is helpful for enhancing collaborative relational learning.

  • The contrastive learning used in variant “DSL-c” attempts to align the social and interaction views, but the irrelevant social connections can mislead the contrastive self-supervision for data augmentation. This observation supports our assumption that social information is inherently noisy for characterizing user preferences.

Parameter Effect Study.

Exploring the influence of key hyperparameters on DSL’s performance, including SSL loss weight, batch size, and the number of graph propagation layers, would be interesting. The results are shown in Figure 2, where the y-axis represents the performance variation ratio compared to the default parameter settings.

  • Effect of SSL regularization weight. The SSL loss weight controls the regularization strength of self-supervision signals. It is clear that a proper weight of SSL loss regularization is beneficial for improving model learning on highly-skewed distributed data. However, the SSL regularization does not always improve model representation. As the SSL loss weight increases, the performance worsens. This is because the model gradient learning is biased towards the strongly regularized SSL signals, which has negative effects on the main optimized objective for recommendation.

  • Effect of batch size. The best model performance is achieved with a batch size of around 2048. The observed performance differences between Epinions and Yelp stem from their diverse social data densities. Epinions data is more sensitive to batch size due to its sparse social connections. Results on Epinions data indicate that a larger batch size helps alleviate the over-fitting issue during model training. However, worse performance is observed with further increasing batch size, possibly due to local optima.

  • Effect of propagation layers #. In GNNs, the number of propagation layers balances the trade-off between informativeness and over-smoothing. When tuning LL from 1 to 4, other parameters are kept at their default settings. A deeper graph neural network is effective in modeling high-order connectivity through cross-layer message passing to generate informative user and item representations. However, stacking too many propagation layers reduces the model capacity by making many different users identical. The resulting over-smoothed embeddings cannot preserve user uniformity with discriminative representations.

Refer to caption
Refer to caption
Refer to caption
Figure 2: We conduct a hyperparameter study of the DSL with respect to i) SSL loss weight for regularization, ii) batch size for training, and iii) # of propagation layers for message passing.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 3: Model robustness study w.r.t data noise and data sparsity, in terms of HR@N and NDCG@N.

4.5 Model Robustness Evaluation (RQ3)

In this section, we investigate the robustness of our DSL against data sparsity and noise for recommendation.

Data Sparsity.

To evaluate the model’s performance on less active users with fewer item interactions, we partitioned the user set into four groups based on their node degrees in the user-item interaction graph 𝒢r\mathcal{G}_{r}, namely (0,5), [5,10), [10,15), and [20, ++\infty). We separately measured the recommendation accuracy for each user group, and the evaluation results are reported in Figure 3. We observed that DSL outperformed the best-performing baseline MHCN in most cases, further validating the effectiveness of our incorporated self-supervision signals for data augmentation under interaction label scarcity.

Data Noise.

To investigate the influence of noisy effects on model performance, we randomly generated different percentages of fake edges (i.e., 10%,20%,30%10\%,20\%,30\%) to create a corrupted interaction graph as noise perturbation. The relative performance degradation with different noise ratios is shown in Figure 3. Our DSL demonstrates great potential in addressing data noise issues compared to competitors. We attribute this superiority to two reasons: 1) Graph structure learning with social relationships as self-supervision signals, which may alleviate the heavy reliance on interaction labels for representation learning. 2) The cross-domain self-augmented learning distills useful information from the interaction view to debias social connections and learn accurate social-aware collaborative relationships.

These findings address the research question raised in RQ3 by demonstrating that our proposed DSL improves the model’s generalization ability over noisy and sparse data.

Refer to caption
Figure 4: Case study on denoising social relations for modeling user-item interaction patterns with sampled socially-connected user pairs.

4.6 Efficiency Analysis (RQ4)

We conduct additional experiments to evaluate the efficiency of our method for model training when DSL competes with baselines. We measure the computational costs (running time) of different methods on an NVIDIA GeForce RTX 3090 and present the training time for each model in Table 5. The training cost of DSL is significantly lower than most of the compared baselines, demonstrating its potential scalability in handling large-scale datasets in real-life recommendation scenarios. While existing social recommenders (e.g., MHCN and SMIN) leverage SSL for data augmentation, blindly maximizing the mutual information between user embeddings may lead to additional computational costs. In DSL, we enhance the social-aware self-supervised learning paradigm with an adaptive denoising module. This simple yet effective framework not only improves the recommendation quality but also shows an advantage in training efficiency.

Data DiffNet NGCF+ MHCN KCGN SMIN Our
Ciao 8.1 8.2 4.92 26.9 7.8 3.2
Epinions 39.1 16.3 9.34 49.4 19.7 6.1
Yelp 692.9 124.6 56.2 132.5 75.3 58.6
Table 5: Model computational cost measured by running time (s).

4.7 Case Study

In this section, we conduct a case study in Figure 4 to qualitatively investigate the effects of our cross-view self-augmented learning framework in denoising social connections for user preference learning. Specifically, we sample two user-user pairs from the Ciao dataset and show the top-33 frequently interacted item categories for each user. From the figure, we observe that the social influence between user 239 and user 227 is identified as weak (i.e., learned lower user relevance weight) with respect to their interaction preference. This is manifested by the fact that they mainly interacted with different item categories, i.e., user 239: Books, Music, and Travel; user 227: Food & Drink, Games, and Household Appliances. On the other hand, the social influence between user 1412 and user 1429 is learned to be strong (i.e., learned higher user relevance weight). Most of their interacted items come from the same categories, i.e., Books, Food & Drink, and Beauty, indicating their similar preferences. This observation aligns with our expectation that DSL can denoise social connections and encode social-aware user interests through meaningful SSL-enhanced representations for recommendation.

5 Conclusion

In this work, we propose a universal denoised self-augmented learning framework that not only incorporates social influence to help understand user preferences but also mitigates noisy effects by identifying social relation bias and denoising cross-view self-supervision. To bridge the gap between social and interaction semantic views, the framework introduces a learnable cross-view alignment to achieve adaptive self-supervised augmentation. Experimental results show that our new DSL leads to significant improvements in recommendation accuracy and robustness compared to existing baselines. Additionally, the component-wise effects are evaluated with ablation study. In future work, we aim to investigate the incorporation of interpretable learning over diverse relations to improve the explainability of denoised self-supervised learners for recommendation. Such incorporation can provide insights into the decision-making process of the social-aware recommender system, enabling users to understand how the system arrives at its recommendation results.

References

  • Cai et al. [2023] Xuheng Cai, Chao Huang, Lianghao Xia, and Xubin Ren. Lightgcl: Simple yet effective graph contrastive learning for recommendation. In International Conference on Learning Representations (ICLR), 2023.
  • Chen et al. [2019a] Chong Chen, Min Zhang, Yiqun Liu, and Shaoping Ma. Social attentional memory network: Modeling aspect-and friend-level differences in recommendation. In WSDM, pages 177–185, 2019.
  • Chen et al. [2019b] Chong Chen, Min Zhang, Chenyang Wang, Weizhi Ma, Minming Li, Yiqun Liu, et al. An efficient adaptive transfer neural network for social-aware recommendation. In SIGIR, pages 225–234, 2019.
  • Chen et al. [2020] Lei Chen, Le Wu, Richang Hong, Kun Zhang, and Meng Wang. Revisiting graph based collaborative filtering: A linear residual graph convolutional network approach. In AAAI, volume 34, pages 27–34, 2020.
  • Chen et al. [2023] Mengru Chen, Chao Huang, Lianghao Xia, Wei Wei, Yong Xu, and Ronghua Luo. Heterogeneous graph contrastive learning for recommendation. In WSDM, pages 544–552, 2023.
  • Epasto and Perozzi [2019] Alessandro Epasto and Bryan Perozzi. Is a single embedding enough? learning node representations that capture multiple social contexts. In WWW, pages 394–404, 2019.
  • Fan et al. [2019] Wenqi Fan, Yao Ma, Qing Li, Yuan He, Eric Zhao, et al. Graph neural networks for social recommendation. In WWW, pages 417–426, 2019.
  • He et al. [2020a] Kaiming He, Haoqi Fan, Yuxin Wu, et al. Momentum contrast for unsupervised visual representation learning. In CVPR, pages 9729–9738, 2020.
  • He et al. [2020b] Xiangnan He, Kuan Deng, et al. Lightgcn: Simplifying and powering graph convolution network for recommendation. In SIGIR, pages 639–648, 2020.
  • Huang et al. [2021] Chao Huang, Huance Xu, Yong Xu, Peng Dai, Lianghao Xia, Mengyin Lu, et al. Knowledge-aware coupled graph neural network for social recommendation. In AAAI, volume 35, pages 4115–4122, 2021.
  • Liu and Liu [2021] Yixin Liu and Pengfei Liu. Simcls: A simple framework for contrastive learning of abstractive summarization. In ACL, 2021.
  • Liu et al. [2019] Ninghao Liu, Qiaoyu Tan, Yuening Li, Hongxia Yang, Jingren Zhou, and Xia Hu. Is a single vector enough? exploring node polysemy for network embedding. In KDD, pages 932–940, 2019.
  • Liu et al. [2022] Zhiwei Liu, Liangwei Yang, Ziwei Fan, Hao Peng, and Philip S Yu. Federated social recommendation with graph neural network. TIST, 13(4):1–24, 2022.
  • Long et al. [2021] Xiaoling Long, Chao Huang, Yong Xu, Huance Xu, Peng Dai, Lianghao Xia, and Liefeng Bo. Social recommendation with self-supervised metagraph informax network. In CIKM, pages 1160–1169, 2021.
  • Mnih and Salakhutdinov [2007] Andriy Mnih and Russ R Salakhutdinov. Probabilistic matrix factorization. NeurIPS, 20, 2007.
  • Rendle et al. [2009] Steffen Rendle, Christoph Freudenthaler, et al. Bpr: Bayesian personalized ranking from implicit feedback. In UAI, pages 452–461, 2009.
  • Song et al. [2019] Weiping Song, Zhiping Xiao, et al. Session-based social recommendation via dynamic graph attention networks. In WSDM, pages 555–563, 2019.
  • Wang et al. [2019] Xiang Wang, Xiangnan He, Meng Wang, Fuli Feng, and Tat-Seng Chua. Neural graph collaborative filtering. In SIGIR, pages 165–174, 2019.
  • Wei et al. [2022a] Chunyu Wei, Jian Liang, Di Liu, and Fei Wang. Contrastive graph structure learning via information bottleneck for recommendation. NeurIPS, 35:20407–20420, 2022.
  • Wei et al. [2022b] Wei Wei, Chao Huang, Lianghao Xia, Yong Xu, Jiashu Zhao, and Dawei Yin. Contrastive meta learning with behavior multiplicity for recommendation. In WSDM, pages 1120–1128, 2022.
  • Wu et al. [2019] Le Wu, Peijie Sun, Yanjie Fu, Richang Hong, Xiting Wang, and Meng Wang. A neural influence diffusion model for social recommendation. In SIGIR, pages 235–244, 2019.
  • Wu et al. [2021] Jiancan Wu, Xiang Wang, Fuli Feng, Xiangnan He, Liang Chen, Jianxun Lian, and Xing Xie. Self-supervised graph learning for recommendation. In SIGIR, pages 726–735, 2021.
  • Yang et al. [2016] Bo Yang, Yu Lei, Jiming Liu, and Wenjie Li. Social collaborative filtering by trust. TPAMI, 39(8):1633–1647, 2016.
  • Yu et al. [2021a] Junliang Yu, Hongzhi Yin, Min Gao, et al. Socially-aware self-supervised tri-training for recommendation. In KDD, pages 2084–2092, 2021.
  • Yu et al. [2021b] Junliang Yu, Hongzhi Yin, Jundong Li, Qinyong Wang, et al. Self-supervised multi-channel hypergraph convolutional network for social recommendation. In WWW, pages 413–424, 2021.
  • Yuan et al. [2022] Xu Yuan, Hongshen Chen, Yonghao Song, Xiaofang Zhao, Zhuoye Ding, Zhen He, and Bo Long. Improving sequential recommendation consistency with self-supervised imitation. In IJCAI, 2022.
  • Zhang et al. [2022] Shengyu Zhang, Lingxiao Yang, Dong Yao, Yujie Lu, et al. Re4: Learning to re-contrast, re-attend, re-construct for multi-interest recommendation. In WWW, pages 2216–2226, 2022.
  • Zhou et al. [2020] Kun Zhou, Hui Wang, Wayne Xin Zhao, Yutao Zhu, Sirui Wang, Fuzheng Zhang, Zhongyuan Wang, and Ji-Rong Wen. S3-rec: Self-supervised learning for sequential recommendation with mutual information maximization. In CIKM, pages 1893–1902, 2020.
  • Zhu et al. [2021] Yanqiao Zhu, Yichen Xu, Feng Yu, Qiang Liu, Shu Wu, et al. Graph contrastive learning with adaptive augmentation. In WWW, pages 2069–2080, 2021.