This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Learning Hierarchical Review Graph Representations for Recommendation

Yong Liu, Susen Yang, Yinan Zhang, Chunyan Miao, Zaiqing Nie, and Juyong Zhang Yong Liu is currently with Joint NTU-UBC Research Centre of Excellence in Active Living for the Elderly (LILY) and Alibaba-NTU Singapore Joint Research Institute, Nanyang Technological University, Singapore 639798. Email: stephenliu@ntu.edu.sg. Yinan Zhang and Chunyan Miao are currently with Joint NTU-UBC Research Centre of Excellence in Active Living for the Elderly (LILY), Alibaba-NTU Singapore Joint Research Institute, and School of Computer Science and Engineering, Nanyang Technological University, Singapore 639798. Email: yinan002@e.ntu.edu.sg, ascymiao@ntu.edu.sg. Susen Yang and Juyong Zhang are currently with School of Mathematical Sciences, University of Science and Technology of China, Hefei, Anhui, China 230052. Email: susen@mail.ustc.edu.cn, juyong@ustc.edu.cn. Zaiqing Nie is currently with the Alibaba Group, Beijing, China. Email: zaiqing.nzq@alibaba-inc.com. Manuscript received xxx, 2020; revised xxx, 2020.
Abstract

The user review data have been demonstrated to be effective in solving different recommendation problems. Previous review-based recommendation methods usually employ sophisticated compositional models, such as Recurrent Neural Networks (RNN) and Convolutional Neural Networks (CNN), to learn semantic representations from the review data for recommendation. However, these methods mainly capture the local dependency between neighbouring words in a word window, and they treat each review equally. Therefore, they may not be effective in capturing the global dependency between words, and tend to be easily biased by noise review information. In this paper, we propose a novel review-based recommendation model, named Review Graph Neural Network (RGNN). Specifically, RGNN builds a specific review graph for each individual user/item, which provides a global view about the user/item properties to help weaken the biases caused by noise review information. A type-aware graph attention mechanism is developed to learn semantic embeddings of words. Moreover, a personalized graph pooling operator is proposed to learn hierarchical representations of the review graph to form the semantic representation for each user/item. We compared RGNN with state-of-the-art review-based recommendation approaches on two real-world datasets. The experimental results indicate that RGNN consistently outperforms baseline methods, in terms of Mean Square Error (MSE).

Index Terms:
Review-based Recommendation, Hierarchical Graph Representation Learning, Graph Neural Networks.

1 Introduction

With the explosive growth of online information and contents, recommendation systems are playing an increasingly important role in various scenarios, e.g., E-commerce websites and online social media platforms. In the last decades, various techniques [1, 2] have been proposed to solve different recommendation problems, for example business rating prediction [3], Point-of-Interests recommendation [4], and diversified item recommendation [5]. Despite many research efforts have been devoted to developing recommendation systems, existing methods may still suffer from the data sparsity problem. To remedy this problem, various types of side information have been incorporated into the recommendation systems, e.g., the item knowledge graph [6], and the users’ reviews on items [7]. In this work, we focus on how to effectively exploit the user review data which usually describe users’ preferences and also include their interests on different aspects of items.

Conventional review-based recommendation methods incorporate topic modeling techniques, e.g., Latent Dirichlet Allocation (LDA) [8] and word embedding model [9], with matrix factorization to jointly learn semantic latent factors for users and items [10, 11, 12, 13, 14]. Recently, the rapid development of deep learning motivates the applications of different neural compositional models, e.g., Convolutional Neural Network (CNN) and Recurrent Neural Network (RNN), to incorporate the review data for recommendation [15, 16, 17, 18, 19]. In addition, there also exists some recent work performing sentiment analysis on the review data to learn aspect-based representations for both users and items [20, 21]. These review-based recommendation methods mainly focus on the rating prediction problem. Following these work, we also study the rating prediction task in this paper.

Despite the successes of existing review-based recommendation methods, they may still suffer from some of the following limitations. Firstly, the methods that only consider document-level co-occurrences of words [10, 11] can only capture the global-level semantic information. They usually ignore the word orders and the local context of words. Secondly, the word embedding and neural compositional models (e.g., CNN and RNN) are effective in capturing the context information from neighbouring words in a word window. However, these methods treat each review equally. As they only rely on the local context of a word, they may be easily biased by the noise review information that are not relevant to the user interests or item properties. Thirdly, most existing methods use the same model to learn the representations of all users and items, without considering the specific characteristics of an individual user/item.

To overcome above limitations, we propose a novel review-based recommendation method, namely Review Graph Neural Network (RGNN). The contributions made in this paper are summarized as follows.

  • We propose to build a directed review graph to describe the reviews of a user/item, where nodes represent the words and edges describe the co-occurrences and order relationships between words. The review graph aggregates the information of all the reviews associated with the user/item to provide a global view about the user/item properties. Then, the reviews with similar contents tend to have more dense connections with each other. Therefore, the review graph can help weaken the biases caused by the noise reviews.

  • We develop the type-aware graph attention (TGAT) mechanism to learn the word embeddings, by aggregating the context information from neighboring words and considering the word orders.

  • Moreover, we also develop the personalized graph pooling (PGP) operator to generate the hierarchical representations of the user/item review graph, considering the user/item-specific properties. PGP is able to select different informative words for different users/items, based on the intuition that the same word may have different semantic informativeness with respect to different users/items.

  • To demonstrate the effectiveness of RGNN, we perform extensive experiments on two real datasets, compared with state-of-the-art review-based recommendation methods.

The remaining parts of this work are organized as follows. In Section 2, we review the most relevant existing work. Then, in Section 3, we introduce the details of the proposed recommendation model. Next, Section 4 reports the experimental results. Finally, Section 5 concludes this work and discusses the potential future work directions.

2 Related Work

This section reviews existing work about review-based recommendation, graph neural networks, and graph pooling.

2.1 Review-based Recommendation Methods

Traditional review-based recommendation methods employ topic modeling techniques on the review data to learn latent feature distributions of users and items. For example, McAuley and Leskovec [10] propose the HFT model that learns the latent factors of users and items from review texts by a LDA model. Guang et al. [12] propose the RMR model that combines the topic model with a mixture of Gaussian models to improve the recommendation accuracy. In [11], Bao et al. employ non-negative matrix factorization to derive topics of each review, and a transform function to align the topic distributions with corresponding latent user/item factors. In [13], Tan et al. propose a rating-boosted method that combines textural reviews with users’ sentiment orientations to learn more accurate topic distributions.

Refer to caption
Figure 1: The overall framework of the proposed RGNN model.

The quick development of deep learning motivates the application of deep learning techniques to extract semantic information from the review data and build the end-to-end recommendation models. For instance, Wang et al. [15] propose the pioneer work that employs stacked denoising autoencoders (SDAE) to learn deep representations from review contents, and probabilistic matrix factorization to model users’ rating behaviors. Zheng et al. [16] propose the DeepCoNN model employing CNN to extract semantic information from the user reviews, and factorization machine to capture the interactions between the user and item features. Recently, the attention mechanism is applied to further improve the recommendation accuracy. In [17], Chen et al. extract the review representation using CNN, and scored each review through an attention mechanism to learn informative user and item representations. In [22], Wu et al. utilize CNN with attention mechanism to highlight relevant semantic information by jointly considering the reviews written by a user and the reviews written to an item. Moreover, Lu et al. [23] propose to jointly learn the user and item information from ratings and reviews by optimizing matrix factorization and an attention-based GRU (i.e., Gated Recurrent Unit) network. Liu et al. [18] propose the dual attention mutual learning between ratings and reviews for recommendation. They use both local and mutual attention of the CNN to jointly learn review representations for improving the model interpretability. Wu et al. [24] utilize three attention networks to learn representations of sentence, review, and user/item in turn, and a graph attention network to model user-item interactions. To incorporate the user/item characteristics, Liu et al. [19] apply a personalized attention model to learn personalized representations of users and items from the review data. In addition, there are some aspect-based recommendation models [20, 21] that apply aspect-based representation learning to automatically detect the most important parts of a review, and jointly learn the aspect, user, and item representations.

2.2 Graph Neural Networks

The graph neural networks (GNNs) extend deep learning techniques to process the graph data. Two main categories of GNN methods are recurrent graph neural networks and graph convolutional neural networks [25].

The pioneer GNN methods are about recurrent graph neural networks, which learn the representation of a node by propagating its neighborhood information in an iterative manner until convergence. For example, in [26], the node’s states are updated by exchanging its neighborhood information recurrently before satisfying the convergence criterion. Gallicchio and Micheli [27] improve the training efficiency of [26] by generalizing the echo state network to graph domains. Moreover, Li et al. [28] propose the gate graph neural networks (GGNN) which employs a gated recurrent unit as the recurrent function. This method no longer needs parameter constraints to ensure convergence.

The graph convolutional neural networks (GCNs) generalize the convolution operation from grid data to graph data. Existing GCN methods usually fall into two groups, i.e., spectral-based methods and spatial-based methods. The spectral-based methods define convolution from the perspective of graph signal processing. For example, Bruna et al. [29] employs the eigen decomposition of the graph Laplacian and defines the convolution in the Fourier domain. To reduce computational complexity, Defferrard et al. [30] approximate the convolutional filters by Chebyshey polynomials of the diagonal matrix of eigenvalues. Then, Kipf and Welling [31] develop a simpler convolutional architecture by the first-order approximation of the CheNet [30]. The spatial-based methods directly operate on the graph and propagate node information along edges. For instance, Micheli [32] proposes to directly sum up a node’s neighborhood information and applies residual connection to memorize information over each layer. Atwood and Towsley [33] treat the graph convolution as a diffusion process, and they assume there is a certain transition probability from a node to its neighbor. As the number of neighbors of nodes varies, Hamilton et al. [34] propose the GraphSage model which samples a fixed number of neighbors for each node and utilizes three different aggregators (e.g., LSTM) to aggregate the neighbors’ feature information. In [35], Veličković et al. propose the graph attention networks (GAT) that assumes the contributions of neighbor nodes to the central node are different. The multi-head attention mechanism is performed to calculate the importance of different neighbors. Moreover, Zhang et al. [36] employ a self-attention mechanism to compute an additional attention score for each attention head. In the recent work [37], Xu et al. identify the graph structures that cannot be distinguished by popular GNN methods, and propose the Graph Isomorphism Network (GIN) which uses a learnable parameter to adjust the weight of the central node.

2.3 Graph Pooling

The pooling operation usually helps provide better performance in graph representation tasks. Earlier pooling methods are mainly based on the deterministic graph clustering algorithms [38, 39], which are less flexible. The recent graph pooling methods can be divided into two groups: spectral and non-spectral methods. The spectral pooling methods [40, 41] employ laplace matrix and eigen-decomposition to capture the graph topology. However, these methods are not scalable to large graph, due to high computation complexity. Thus, a lot of research attentions have been attracted to develop the non-spectral pooling methods, which can be further divided into global and hierarchical pooling methods. The global pooling methods [42, 43] pool the representations of all nodes at one step. For example, Vinyals et al. [42] propose the Set2Set model that aggregates nodes information by combining LSTM with a content-based attention mechanism. Zhang et al. [43] develop the SortPool model that first sorts nodes according to their features and then concatenates the embedding of sorted nodes to summarize the graph. Differing from the global pooling methods, the hierarchical pooling methods aim to learn hierarchical graph representations. For example, Ying et al. [44] propose the first end-to-end differentiable pooling operator, named DiffPool, which learns a dense soft clustering assignment matrix of nodes by GNN. The gPool [45, 46] and SAGPool [47] pooling operators utilize a node-scoring procedure to select a set of top ranked nodes to form an induced graph from original graph. Moreover, Diehl [48] proposes a hard pooling method called EdgePool, which is developed based on edge contraction. In [49], Ranjan et al. utilize a self-attention mechanism to capture the importance of each node. This method learns a sparse soft cluster assignment for nodes to pool the graph. In [50], Zhang et al. propose a non-parametric pooling operation using a structure learning mechanism with sparse attention to learn a refined graph structure that preserves the key substructure in original graph.

3 The Proposed Recommendation Model

This work focuses on exploiting users’ review data for recommendation. For each user uu, the set of her reviews is denoted by 𝒮u={su1,su2,,suMu}\mathcal{S}_{u}=\{s_{u}^{1},s_{u}^{2},\cdots,s_{u}^{M_{u}}\}, where suts_{u}^{t} denotes the tt-th review of user uu, and MuM_{u} denotes the number of reviews generated by the user uu. Similarly, let 𝒯i={τi1,τi2,,τiNi}\mathcal{T}_{i}=\{\tau_{i}^{1},\tau_{i}^{2},\cdots,\tau_{i}^{N_{i}}\} denote the reviews of each item ii, where NiN_{i} denotes the number of reviews generated on the item ii, and τit\tau_{i}^{t} denotes the tt-th review of item ii. Note that each review may consist of multiple text sentences. In addition, we denote the observed rating given by the user uu on the item ii by yuiy_{ui}. For a word xx, we denote its word embedding by 𝐞x1×d0\mathbf{e}_{x}\in\mathbb{R}^{1\times d_{0}}, where d0d_{0} denotes the dimensionality of the latent semantic space. Moreover, we use 𝐞u1×d1\mathbf{e}_{u}\in\mathbb{R}^{1\times d_{1}} and 𝐞i1×d1\mathbf{e}_{i}\in\mathbb{R}^{1\times d_{1}} to denote the embedding of user uu and item ii, respectively, where d1d_{1} is the dimensionality of latent rating space.

The overall framework of the proposed RGNN model is summarized in Figure 1. As shown in Figure 1, RGNN contains three components, i.e., the User-Net, Item-Net, and a prediction layer. The User-Net and Item-Net are two parallel networks used to learn semantic representations for users and items, respectively. The prediction layer is developed based on factorization machine (FM) [51], which aims to predict the rating scores by considering the interactions between learned semantic features. Both User-Net and Item-Net consist of the following three modules:

  • The review graph construction module builds the review graph for each user and item, considering the review words as nodes in the graph and capturing the co-occurrence and order relationships between words.

  • The graph representation learning module applies a type-aware graph attention network (TGAT) and personalized graph pooling operator (PGP) to extract the user/item-specific hierarchical representations from the entire review graph.

  • Moreover, the feature fusion module integrates the hierarchical representations of the user/item review graph and the user/item embedding to form the semantic representation for the user/item.

As User-Net and Item-Net only differ in their inputs, we only introduce the details of User-Net in the following sections. The same procedure is applied to the Item-Net.

3.1 Review Graph Construction

Various graph based methods have been proposed to process the text data [52], which aim to capture the inherent dependence between words. Here, we utilize the graph-of-words method [53] to construct the review graph with preserving the word orders. For the reviews 𝒮u\mathcal{S}_{u} of the user uu, we firstly apply the phrase pre-processing techniques such as tokenization and text cleaning to select keywords of each review in 𝒮u\mathcal{S}_{u}. Then, all reviews in 𝒮u\mathcal{S}_{u} can be encoded into an un-weighted and directed graph, where nodes represent words and edges describe the co-occurrences between words within a sliding window with fixed size ω\omega. Note that we define the word windows considering the positions of selected keywords in the original text.

Intuitively, the order information of words in the text may help capture the sentiment information [54]. For example, the phrases “not really good” and “really not good” convey different levels of negative sentiment information, which is only caused by the word orders. To preserve the orders of words in the review, we define three types of edges in the review graph: forward type, backward type, and self-loop type (respectively denoted by efe_{f}, ebe_{b}, and ese_{s}). We denote the set of edge type by ={ef,eb,es}\mathcal{R}=\{e_{f},e_{b},e_{s}\}. In a review sut𝒮us_{u}^{t}\in\mathcal{S}_{u}, if the selected keyword (i.e., node) x2x_{2} appears before the selected keyword x1x_{1} and the distance between x1x_{1} and x2x_{2} in the original text of suts_{u}^{t} is less than ω\omega, we construct an edge E(x2,x1)E(x_{2},x_{1}) from x2x_{2} to x1x_{1} and set the type of edge E(x2,x1)E(x_{2},x_{1}) as efe_{f}. Moreover, we also construct an edge E(x1,x2)E(x_{1},x_{2}) from x1x_{1} to x2x_{2} and set the type of E(x1,x2)E(x_{1},x_{2}) as ebe_{b}. In the rare situation where the word x2x_{2} may appear both before and after the word x1x_{1}, we randomly set the type of edge E(x1,x2)E(x_{1},x_{2}) as efe_{f} or ebe_{b}. In the experiments, we empirically set ω\omega to 3. To consider the information of the word itself (e.g., x1x_{1}), we add a self-loop edge at each word (e.g., E(x1,x1)E(x_{1},x_{1})) and set the edge type as ese_{s}. This trick also helps define the attention score in Eq. (1).

Figure 2 shows an example about constructing the review graph for the review “My son likes this beautiful looking football”, where the words “son”, “likes”, “beautiful”, “looking”, and “football” are selected as keywords to build the review graph. In the original review text, the distance between “son” and “likes” is 1 which is less than ω\omega. Thus, we construct an edge from “son” to “likes” and set the edge type as efe_{f}. Moreover, we also construct an edge from “likes” to “son” and set the edge type as ebe_{b}. However, the distance between “son” and “beautiful” in original text is 3, which is not smaller than the windows size (ω=3\omega=3). Therefore, as shown in Figure 2, there is no connection between the selected keywords “son” and “beautiful”.

For each user uu, we denote her review graph by 𝒢u={𝒳u,u}\mathcal{G}_{u}=\{\mathcal{X}_{u},\mathcal{E}_{u}\}, where 𝒳u\mathcal{X}_{u} denotes the set of nodes (i.e., selected keywords) and u\mathcal{E}_{u} denotes the set of node-edge-node triples (xh,er,xt)(x_{h},e_{r},x_{t}). Here, ere_{r} is type of the edge connecting from node xtx_{t} to xhx_{h}, which is one of the three edge types defined above. Similarly, we can build the review graph 𝒢i={𝒳i,i}\mathcal{G}_{i}=\{\mathcal{X}_{i},\mathcal{E}_{i}\} for each item ii.

Refer to caption
Figure 2: An example of constructing graph for the review “My son likes this beautiful looking football”. The words “son”, “likes”, “beautiful”, “looking”, and “football” are selected as nodes. Empirically, we set the window size ω\omega to 3 for edge building.
Refer to caption
Figure 3: The procedure of hierarchical graph representation learning.

3.2 Hierarchical Graph Representation Learning

After completing the construction of review graph 𝒢u\mathcal{G}_{u}, we employ the representation learning procedure as shown in Figure 3 to learn hierarchical representations of 𝒢u\mathcal{G}_{u}, by repeating the TGAT and PGP operations multiple times. In this section, we introduce the details of the operations in the \ell-th representation learning layer, where the input graph is 𝒢u={𝒳u,u}\mathcal{G}_{u}^{\ell}=\{\mathcal{X}_{u}^{\ell},\mathcal{E}_{u}^{\ell}\}. Suppose 𝒢u\mathcal{G}_{u}^{\ell} has KuK_{u}^{\ell} nodes. Then, we also use an adjacency matrix 𝐀uKu×Ku\mathbf{A}^{\ell}_{u}\in\mathbb{R}^{K_{u}^{\ell}\times K_{u}^{\ell}} and a feature matrix 𝐗uKu×d2\mathbf{X}_{u}^{\ell}\in\mathbb{R}^{K_{u}^{\ell}\times d_{2}^{\ell}} to describe 𝒢u\mathcal{G}_{u}^{\ell}, where each row in 𝐗u(x1,:)\mathbf{X}_{u}^{\ell}(x_{1},:) denotes the input feature of the node x1x_{1}, and d2d_{2}^{\ell} denotes the dimensionality of the input feature. Node that the features of nodes in the review graph are initialized by the embeddings of corresponding words. Thus, at the first layer, d2d_{2}^{\ell} is equal to d0d_{0}. In 𝒢u\mathcal{G}_{u}^{\ell}, if there exists an edge between two nodes x1x_{1} and x2x_{2}, we set 𝐀u[x1,x2]\mathbf{A}^{\ell}_{u}[x_{1},x_{2}] to 1, and set to 0, otherwise.

3.2.1 Type-aware Graph Attention Network

In the reviews, the semantic meaning of a word can be enriched by its neighboring words, which can be considered as the dependency between a node and its neighboring nodes in the review graph. At the \ell-th representation learning layer, for a node xhx_{h} in the graph 𝒢u\mathcal{G}_{u}^{\ell}, we denote its neighborhood in the graph by 𝒩h={xt|(xh,er,xt)u}\mathcal{N}_{h}^{\ell}=\{x_{t}|(x_{h},e_{r},x_{t})\in\mathcal{E}_{u}^{\ell}\}. Note that 𝒩h\mathcal{N}_{h}^{\ell} also includes the node xhx_{h}, because there exists an self-loop edge at each node in the review graph. A type-aware attention mechanism is developed to effectively aggregate the information of neighboring words, considering the types of connecting edges. Following [35], we define the attention weight a(xh,er,xt)a^{\ell}(x_{h},e_{r},x_{t}) of a neighboring node xt𝒩hx_{t}\in\mathcal{N}_{h}^{\ell} as,

a(xh,er,xt)=exp[π(xh,er,xt)]xt~𝒩hexp[π(xh,er~,xt~)],a^{\ell}(x_{h},e_{r},x_{t})=\frac{\exp[\pi^{\ell}(x_{h},e_{r},x_{t})]}{\sum_{x_{\tilde{t}}\in\mathcal{N}_{h}^{\ell}}\exp[\pi^{\ell}(x_{h},e_{\tilde{r}},x_{\tilde{t}})]}, (1)

where π(xh,er,xt)\pi^{\ell}(x_{h},e_{r},x_{t}) is implemented by the following relational attention mechanism,

π(xh,er,xt)=LeakyReLU[(𝐱h𝐖1)(𝐱t𝐖2+𝐞r𝐖3)].\pi^{\ell}(x_{h},e_{r},x_{t})=\mbox{LeakyReLU}\big{[}(\mathbf{x}^{\ell}_{h}\mathbf{W}^{\ell}_{1})(\mathbf{x}^{\ell}_{t}\mathbf{W}^{\ell}_{2}+\mathbf{e}^{\ell}_{r}\mathbf{W}^{\ell}_{3})^{\top}\big{]}. (2)

Here, 𝐱h,𝐱t1×d2\mathbf{x}_{h}^{\ell},\mathbf{x}_{t}^{\ell}\in\mathbb{R}^{1\times d_{2}^{\ell}} are the input embeddings of xhx_{h} and xtx_{t} at the \ell-th layer. 𝐞r1×d2\mathbf{e}_{r}^{\ell}\in\mathbb{R}^{1\times d_{2}^{\ell}} is the embeddings of ere_{r} at the \ell-th layer. 𝐖1,𝐖2,𝐖3d2×d2\mathbf{W}_{1}^{\ell},\mathbf{W}_{2}^{\ell},\mathbf{W}_{3}^{\ell}\in\mathbb{R}^{d_{2}^{\ell}\times d_{2}} are three different weight matrices used for the linear transformations of the node and edge embeddings, and d2d_{2} denotes the dimensionality of the hidden space of the graph representation learning layers. The attention weights indicate which neighboring node should contribute more to the central node, when aggregating neighborhood information. Then, we can compute the linear combination of neighbors as follows,

𝐱¯h=xt𝒩ha(xh,er,xt)𝐱t𝐖4,\bar{\mathbf{x}}_{h}^{\ell}=\sum_{x_{t}\in\mathcal{N}_{h}^{\ell}}a^{\ell}(x_{h},e_{r},x_{t})\mathbf{x}^{\ell}_{t}\mathbf{W}_{4}^{\ell}, (3)

where 𝐖4d2×d2\mathbf{W}_{4}^{\ell}\in\mathbb{R}^{d_{2}^{\ell}\times d_{2}} denotes the transform matrix. A feature matrix 𝐗¯uKu×d2\bar{\mathbf{X}}_{u}^{\ell}\in\mathbb{R}^{K_{u}^{\ell}\times d_{2}} is used to denote the output of TGAT. Each row of 𝐗¯u\bar{\mathbf{X}}_{u}^{\ell} corresponds to the feature of a node xx in 𝒢u\mathcal{G}_{u}^{\ell}, computed by Eq. (3). Then, 𝐗¯u\bar{\mathbf{X}}_{u}^{\ell} will be used as the input of the PGP operation at the \ell-th representation learning layer.

Compared with the methods using CNN to extract semantic information, the proposed TGAT operation can not only capture the global contextual word information from all the reviews of a user/item, but also specify varying importance weights to different contextual words. In addition, we explicitly incorporate the edge type ere_{r} into the attention mechanism, which encodes word order information during aggregating neighboring contextual word information.

3.2.2 Personalized Graph Pooling

The graph pooling operation plays an important role in learning graph representations. Various hierarchical graph pooling methods have been proposed to adaptively select a subset of important nodes to form a smaller sub-graph [45, 46, 47, 49, 50]. However, these methods ignore the user/item-specific properties of an individual user/item, thus they are not suitable to learn the review graph representation for recommendation. In practice, different users may emphasize different aspects of items. For example, some users are most interested in the item price, while some other users may be more interested in the item quality. Similarly, the most important properties of different items may also be different. Therefore, it is intuitively that the same word may have different importance to different users/items. Motivated by this intuition, we propose the personalized graph pooling operator to capture the important nodes (i.e., words) for each user/item.

For the user uu, we first use a Multilayer Perceptron (MLP) to generate her personalized measurement vector 𝜽ul\bm{\theta}_{u}^{l} for measuring the importance of nodes at the \ell-th representation learning layer,

𝜽u=ReLU(𝐞u𝐖5+𝐛5),\bm{\theta}_{u}^{\ell}=\mbox{ReLU}\big{(}\mathbf{e}_{u}\mathbf{W}_{5}^{\ell}+\mathbf{b}_{5}^{\ell}\big{)}, (4)

where 𝐖5d1×d2\mathbf{W}_{5}^{\ell}\in\mathbb{R}^{d_{1}\times d_{2}} and 𝐛51×d2\mathbf{b}_{5}^{\ell}\in\mathbb{R}^{1\times d_{2}} are the weight matrix and bias term of MLP at the \ell-th pooling layer. We adapt graph pooling to the recommendation scenario studied in this work, and define the following importance score βh\beta_{h}^{\ell} for each node xh𝒳ux_{h}\in\mathcal{X}_{u}^{\ell} as,

βh=|𝐱¯h𝜽u|+1|𝒩h|xt𝒩h|(𝐱¯h𝐱¯t)𝐰|,\beta_{h}^{\ell}=\big{|}\bar{\mathbf{x}}_{h}^{\ell}{\bm{\theta}_{u}^{\ell}}^{\top}\big{|}+\frac{1}{|\mathcal{N}^{\ell}_{h}|}\sum_{x_{t}\in\mathcal{N}^{\ell}_{h}}\big{|}(\bar{\mathbf{x}}^{\ell}_{h}-\bar{\mathbf{x}}^{\ell}_{t}){\mathbf{w}^{\ell}}^{\top}\big{|}, (5)

where 𝐰1×d2\mathbf{w}^{\ell}\in\mathbb{R}^{1\times d_{2}} is a learnable parameter. In Eq. (5), the first part captures the importance of node xhx_{h} to the user uu. The second part describes the feature diversity of xhx_{h}’s neighborhood at the \ell-th layer. It is developed based on the criterion that a node xhx_{h} can be removed in the pooled graph with negligible information loss, if xhx_{h} can be well reconstructed by its neighboring nodes in 𝒢u\mathcal{G}_{u}^{\ell}. Moreover, 𝐰\mathbf{w}^{\ell} can also help balance the contributions of these two parts in the \ell-th layer.

Then, the score vector 𝜷=[β1,β2,,βKu]\bm{\beta}^{\ell}=[\beta_{1}^{\ell},\beta_{2}^{\ell},\cdots,\beta_{K_{u}^{\ell}}^{\ell}]^{\top} is multiplied with the feature matrix 𝐗¯ul\bar{\mathbf{X}}_{u}^{l} to obtain the following filtered feature matrix,

𝐗^u=𝐃𝐗¯u,\widehat{\mathbf{X}}_{u}^{\ell}=\mathbf{D}^{\ell}\bar{\mathbf{X}}_{u}^{\ell}, (6)

where 𝐃Ku×Ku\mathbf{D}^{\ell}\in\mathbb{R}^{K_{u}^{\ell}\times K_{u}^{\ell}} is a diagonal matrix. The diagonal element Dhh=σ(βh)D_{hh}^{\ell}=\sigma(\beta_{h}^{\ell}), and σ()\sigma(\cdot) denotes the sigmoid function. Then, we can obtain the adjacency matrix 𝐀u+1\mathbf{A}_{u}^{\ell+1} and feature matrix 𝐗u+1\mathbf{X}_{u}^{\ell+1} of the pooled graph 𝒢u+1\mathcal{G}_{u}^{\ell+1} as follows,

idx=rank(𝜷,Ku+1),\displaystyle\mbox{idx}=\mbox{rank}(\bm{\beta}^{\ell},K_{u}^{\ell+1}), (7)
𝐀u+1=𝐀u\displaystyle\mathbf{A}_{u}^{\ell+1}=\mathbf{A}_{u}^{\ell} (idx,idx),𝐗u+1=𝐗^u(idx,:),\displaystyle(\mbox{idx},\mbox{idx}),\mathbf{X}_{u}^{\ell+1}=\widehat{\mathbf{X}}^{\ell}_{u}(\mbox{idx},:), (8)

where rank()\mbox{rank}(\cdot) returns the indices of the top-Ku+1K_{u}^{\ell+1} values in 𝜷\bm{\beta}^{\ell}, 𝐀u(idx,idx)\mathbf{A}^{\ell}_{u}(\mbox{idx},\mbox{idx}) is the row-wise and column-wise indexed adjacency matrix, and 𝐗^u(idx,:)\widehat{\mathbf{X}}^{\ell}_{u}(\mbox{idx},:) is the row-wise indexed feature matrix. Then, we have 𝐀u+1\mathbf{A}_{u}^{\ell+1} and 𝐗u+1\mathbf{X}_{u}^{\ell+1} which are new adjacency matrix and corresponding feature matrix of a smaller graph 𝒢u+1\mathcal{G}_{u}^{\ell+1}. In this work, we empirically set Ku+1=αKuK_{u}^{\ell+1}=\lceil\alpha K_{u}^{\ell}\rceil, where α\alpha is the pooling ratio, \lceil\cdot\rceil is the ceiling function.

3.3 Feature Fusion

After repeating the TGAT and PGP operations LL times, we can obtain multiple pooled sub-graphs with different size, i.e., 𝒢u2,𝒢u3,,𝒢uL+1\mathcal{G}_{u}^{2},\mathcal{G}_{u}^{3},\cdots,\mathcal{G}_{u}^{L+1}. This hierarchical pooling operation extracts the coarser-grained information of review graph layer-by-layer, which can be considered as different level representations of reviews, e.g., word-level, sentence-level, and review-level. To integrate these sub-graph information, we first use the max-pooling and MLP to generate representation of each sub-graph 𝒢u\mathcal{G}_{u}^{\ell}, where 2L+12\leq\ell\leq L+1, as follows:

𝐠^u=maxt=1d2𝐗u(:,t),𝐠u=ReLU(𝐠^u𝐖6+𝐛6),\displaystyle\hat{\mathbf{g}}_{u}^{\ell}=\mathop{\mbox{max}}\limits_{t=1}^{d_{2}}\mathbf{X}_{u}^{\ell}(:,t),~{}~{}\mathbf{g}_{u}^{\ell}=\mbox{ReLU}(\hat{\mathbf{g}}_{u}^{\ell}\mathbf{W}_{6}^{\ell}+\mathbf{b}_{6}^{\ell}), (9)

where 𝐖6d2×d1\mathbf{W}_{6}^{\ell}\in\mathbb{R}^{d_{2}\times d_{1}} and 𝐛61×d1\mathbf{b}_{6}^{\ell}\in\mathbb{R}^{1\times d_{1}} are the weight matrix and bias vector. 𝐠u\mathbf{g}_{u}^{\ell} denotes the representation of the sub-graph generated at the \ell-th layer. Then, we concatenate the non-linear transform of the user embedding after a MLP layer and the representations of sub-graphs at all layers to generate the semantic representation 𝐩u\mathbf{p}_{u} of the user uu as follows,

𝐦u\displaystyle\mathbf{m}_{u} =ReLU(𝐞u𝐖7+𝐛7),\displaystyle=\mbox{ReLU}(\mathbf{e}_{u}\mathbf{W}_{7}+\mathbf{b}_{7}),
𝐩u\displaystyle\mathbf{p}_{u} =𝐦u𝐠u2𝐠u3𝐠uL+1,\displaystyle=\mathbf{m}_{u}\oplus\mathbf{g}_{u}^{2}\oplus\mathbf{g}_{u}^{3}\oplus\cdots\oplus\mathbf{g}_{u}^{L+1}, (10)

where \oplus is the concatenating operation, 𝐖7d1×d1\mathbf{W}_{7}\in\mathbb{R}^{d_{1}\times d_{1}} and 𝐛71×d1\mathbf{b}_{7}\in\mathbb{R}^{1\times d_{1}} are the weight matrix and bias vector. Similarly, we can get the semantic representation of the item ii as 𝐪i\mathbf{q}_{i}.

3.4 Rating Prediction

A factorization machine (FM) layer is used to predict the user’s preference on an item, considering the higher order interactions between the user and item semantic features. Specifically, we first concatenate the user and item semantic representations as 𝐳=𝐩u𝐪i\mathbf{z}=\mathbf{p}_{u}\oplus\mathbf{q}_{i}, and the prediction y^ui\hat{y}_{ui} of the user’s preference can be defined as follows,

y^ui=b0+bu+bi+𝐳𝐰+i=1dj=i+1d𝐯i,𝐯jzizj,\hat{y}_{ui}=b_{0}+b_{u}+b_{i}+\mathbf{z}\mathbf{w}^{\top}+\sum_{i=1}^{d^{\prime}}\sum_{j=i+1}^{d^{\prime}}\langle\mathbf{v}_{i},\mathbf{v}_{j}\rangle z_{i}z_{j}, (11)

where b0b_{0}, bub_{u} and bib_{i} are the global bias, user bias, and item bias. 𝐰1×d\mathbf{w}\in\mathbb{R}^{1\times d^{\prime}} is the coefficient vector, and d=2(L+1)d1d^{\prime}=2*(L+1)*d_{1}. 𝐯i,𝐯j1×k\mathbf{v}_{i},\mathbf{v}_{j}\in\mathbb{R}^{1\times k} are the latent factors associated with ii-th and jj-th dimension of 𝐳\mathbf{z}. Empirically, we set k=d2k=d_{2} in this work. ziz_{i} is the value of the ii-th dimension of 𝐳\mathbf{z}. In this work, we focus on the rating prediction problem. Thus, the model parameters 𝚯\mathbf{\Theta} of RGNN can be learned by solving the following optimization problem,

min𝚯(u,i)𝒟(yuiy^ui)2+λ𝚯F2,\min_{\mathbf{\Theta}}\sum_{(u,i)\in\mathcal{D}}\big{(}y_{ui}-\hat{y}_{ui}\big{)}^{2}+\lambda\big{\|}\mathbf{\Theta}\big{\|}_{F}^{2}, (12)

where λ\lambda is the regularization parameter, 𝒟\mathcal{D} denotes the set of user-item pairs used to update the model parameters. The entire framework can be effectively trained by the end-to-end back propagation algorithm.

3.5 Discussions

In this work, we build a review graph to describe all the reviews of a user/item. If a word appears in multiple review sentences, its neighbors in review graph aggregate all its neighboring words (within a word window) in these sentences. Then, TGAT is proposed to learn the word embedding, considering its global context words among all the relevant reviews. However, the CNN/RNN-based methods separately aggregate the local neighboring words of a word in different review sentences. Moreover, TGAT employs attention mechanism to automatically assign different importance weights to neighboring words. By performing multiple times of PGP operations, the proposed model can capture the long-term dependency between selected words.

Suppose the average number of selected keywords for each user/item is NwN_{w}. The time complexity of constructing the review graph is O(ωNw)O\big{(}\omega N_{w}\big{)}. In practice, the review graph construction can be performed offline. The time complexity for RGNN to obtain the hierarchical review representation is O(=1LαNwd02)O\big{(}\sum_{\ell=1}^{L}\alpha^{\ell}N_{w}d_{0}^{2}\big{)}, where LL denotes the number of layers, α\alpha and d0d_{0} denote the pooling ratio and word embedding dimension. The total training time complexity of RGNN is O(LNwd02)O\big{(}LN_{w}d_{0}^{2}\big{)}. Moreover, the time complexity of using CNNs to obtain the review representation is O(MSNwd0)O\big{(}MSN_{w}d_{0}\big{)}, where MM and SS denote the number of convolutional kernels and the convolutional kernel size, respectively. Thus, compared with the CNN based methods, the proposed RGNN model is not complicate or can be even more efficient.

4 Experiments

We empirically perform extensive experiments on real datasets to demonstrate the effectiveness of the proposed RGNN method, comparing with state-of-the-art review-based recommendation methods.

4.1 Experimental Setting

4.1.1 Datasets

The experiments are performed on the Amazon review dataset [55] and Yelp dataset111https://www.yelp.com/dataset, which have been widely used for recommendation research. For Amazon review dataset, we choose the following 5-core review subsets for evaluation: “Musical Instruments”, “Office Products”, “Toys and Games”, “Beauty”, and “Video Games” (respectively denoted by Music, Office, Toys, Beauty, and Games). For Yelp dataset, we only keep the users and items that have at least 10 reviews for experiments. Table I summarizes the details of these experimental datasets.

TABLE I: The statistics of the experimental datasets.
Dataset # Users # Items # Ratings Density
Music 1,429 900 10,261 0.80%
Office 4,905 2,420 53,228 0.45%
Toys 19,412 11,924 167,597 0.07%
Beauty 22,363 12,101 198,502 0.07%
Games 24,303 10,672 231,577 0.09%
Yelp 126,084 65,786 3,519,533 0.04%
TABLE II: The performances of different recommendation algorithms evaluated by MSE. The best results are in bold faces and the second best results are underlined. \blacktriangle denotes the average relative improvements achieved by RGNN over baselines on all datasets. indicates RGNN significantly outperforms all baselines with p<0.05p<0.05 using student t-test on the dataset.
Dataset PMF SVD++ CDL DeepCoNN ANR NARRE CARL NRPA RMG DAML RGNN
Music 1.8783 0.7952 1.2987 0.7909 0.7825 0.7688 0.7632 0.7658 0.7560 0.7401 0.7319
Office 0.9635 0.7213 0.8763 0.7315 0.7237 0.7266 0.7193 0.7343 0.7348 0.7164 0.7125
Toys 1.6091 0.8276 1.2479 0.8073 0.7974 0.7912 0.8248 0.8100 0.8435 0.7909 0.7786
Games 1.5260 1.2081 1.6002 1.1234 1.1038 1.1120 1.1308 1.1259 1.1496 1.1086 1.0996
Beauty 2.7077 1.2129 1.7726 1.2210 1.2021 1.1997 1.2250 1.2034 1.2419 1.2175 1.1885
Yelp 1.4217 1.2973 1.4042 1.2719 1.2708 1.2675 1.3199 1.2721 1.2981 1.2700 1.2645
\blacktriangle Improvements 38.97% 4.77% 29.02% 3.16% 2.06% 1.77% 3.43% 2.57% 4.19% 1.14% -

4.1.2 Setup and Metric

For each dataset, we randomly choose 20% of the user-item review pairs (denoted by 𝒟test\mathcal{D}_{test}) for evaluating the model performance in the testing phase, and the remaining 80% of the review pairs (denoted by 𝒟train\mathcal{D}_{train}) are used in the training phase. Then, 10% of the data in 𝒟train\mathcal{D}_{train} are held out as the validation set to choose the model hyper-parameters, and the remaining 90% of the data in 𝒟train\mathcal{D}_{train} are used to update the model parameters. As this work focuses on the rating prediction task, the performances of the recommendation algorithms are typically evaluated by Mean Square Error (MSE), which has been widely used in previous studies [16, 21, 22, 19]. The definition of MSE is as follows:

MSE=1|𝒟test|(u,i)𝒟test(yuiy^ui)2,\displaystyle MSE=\frac{1}{|\mathcal{D}_{test}|}\sum_{(u,i)\in\mathcal{D}_{test}}(y_{ui}-\hat{y}_{ui})^{2}, (13)

where 𝒟test\mathcal{D}_{test} denotes the set of all tested user-item rating pairs. From the definition in Eq. 13, we can note that lower MSE indicates better performances.

4.1.3 Baseline Methods

We compare RGNN with the following recommendation methods:

  • PMF [56]: This is the probabilistic matrix factorization model, which is a classical collaborative filtering based rating prediction method.

  • SVD++ [57]: This is a classic matrix factorization method that exploits both the user’s explicit preferences on items and the influences of the user’s historical items on the target item.

  • CDL [15]: This is a hierarchical Bayesian model that employs SDAE for learning features from the content information and collaborative filtering for modeling the rating behaviors.

  • DeepCoNN [16]: This method contains two parallel networks, which focus on modeling the user behaviors and learning the item properties from the review data.

  • ANR [21]: This is aspect-based neural recommendation model that learns aspect-based representations for the user and item by an attention-based module. Moreover, the co-attention mechanism is applied to the user and item importance at the aspect-level.

  • NARRE [17]: This method uses an attention mechanism to model the importance of reviews and a neural regression model with review-level explanation for rating prediction.

  • CARL [22]: This is a context-aware representation learning model for rating prediction, which uses convolution operation and attention mechanism for review-based feature learning and factorization machine for modeling high-order feature interactions.

  • NRPA [19]: This method uses a review encoder to learn the review representation, and a user/item encoder with personalized attention mechanism to learn user/item representations from reviews.

  • RMG [24]: This method utilizes a multi-view learning framework to incorporate information from the review-content view and user-item interaction view for recommendation.

  • DAML [18]: This method employs CNN with local and mutual attention mechanism to learn the review features and improve the interpretability of the recommendation model.

TABLE III: Performances of RGNN with respect to different settings of LL.
Music Office Toys Games Beauty Yelp
L=1 0.7572 0.7130 0.7830 1.1021 1.1964 1.2671
L=2 0.7319 0.7125 0.7786 1.1025 1.1946 1.2658
L=3 0.7557 0.7145 0.7825 1.0996 1.1903 1.2645
L=4 0.7557 0.7149 0.7802 1.1017 1.1885 1.2684

4.1.4 Parameter Settings

In this work, we implement the proposed RGNN model by Pytorch and choose Adam [58] as the optimization method to learn the model parameters. The dimensionality of the semantic space d0d_{0}, the rating space d1d_{1}, and the hidden space of the graph representation learning layers d2d_{2} are chosen from [4,8,16,32,48,64,128][4,8,16,32,48,64,128]. The learning rate γ\gamma is varied in [0.0005,0.001,0.005][0.0005,0.001,0.005], and regularization parameter λ\lambda is varied in [0.001,0.01,0.05,0.1,1.0][0.001,0.01,0.05,0.1,1.0]. The batch size is chosen in [64,128,256][64,128,256]. Moreover, the number of layers LL is varied in [1,2,3,4][1,2,3,4], and the pooling ratio α\alpha is set to 0.5.

The hyper-parameters of baseline methods are set following original papers. In PMF and SVD++, the dimension of latent embeddings are chosen from {16,32,48,64,128}\{16,32,48,64,128\}. The learning rate is varied in {0.0001,0.001,0.01}\{0.0001,0.001,0.01\}, and the batch size is chosen from {64,128,256,512}\{64,128,256,512\}. For CDL, we set a=1, b=0.01, K=50, and choose λu{\lambda}_{u}, λv{\lambda}_{v}, λn{\lambda}_{n}, λw{\lambda}_{w} from {0.01,0.1,1,10,100}\{0.01,0.1,1,10,100\}. The number of hidden factors is selected from {50,100,200,300}\{50,100,200,300\}, and dropout ratio is varied in {0.1,0.3,0.5,0.7,0.9}\{0.1,0.3,0.5,0.7,0.9\}. For DeepCoNN, we empirically vary the number of latent factors (i.e., |𝐱u||\mathbf{x}_{u}| and |𝐲i||\mathbf{y}_{i}|) from 5 to 100 and the number of convolutional kernels from 10 to 400. The window size is set to 3, and the learning rate is chosen from {0.0001,0.001,0.002,0.01}\{0.0001,0.001,0.002,0.01\}. In ANR, the dimensionality of latent representations is chosen from {64,128,256}\{64,128,256\}. The window size is set to 3. The number of aspect is selected from {2,4,6,8}\{2,4,6,8\}. Moreover, the dimension of aspect embeddings is chosen from {5,10,15,25,50}\{5,10,15,25,50\}, and the dropout ratio is selected from {0.1,0.3,0.5,0.7,0.9}\{0.1,0.3,0.5,0.7,0.9\}. For NARRE, the learning rate is selected from {0.0001,0.001,0.002,0.01,0.05}\{0.0001,0.001,0.002,0.01,0.05\}. The dropout ratio is chosen from {0.1,0.3,0.5,0.7,0.9}\{0.1,0.3,0.5,0.7,0.9\}. The batch size is selected from {50,100,150}\{50,100,150\}, and the latent factor number is varied in {16,32,64,128}\{16,32,64,128\}. The word window size is to 3. For CARL, we select the dimensionality of latent embeddings from {15,50,100,200,300}\{15,50,100,200,300\} and the batch size from {100,200}\{100,200\}. The learning rate is chosen from {0.0001,0.001,0.01}\{0.0001,0.001,0.01\}, and the dropout ratio is chosen from {0.1,0.3,0.5,0.7,0.9}\{0.1,0.3,0.5,0.7,0.9\}. The word window size is also set to 3. In NRPA, the dimensionality of embeddings is chosen from {16,32,64,128}\{16,32,64,128\}. The number of filters in CNN and the dimension of attention vectors are chosen from {20,40,80,100}\{20,40,80,100\}. The word window size is set to 3. For RMG, the number of CNN filters is chosen from {50,100,150,200}\{50,100,150,200\}, and the dimensionality of embeddings is chosen from {50,100,150,200}\{50,100,150,200\}. The batch size is varied in {20,50,100,200}\{20,50,100,200\}, and the dropout ratio is selected from {0.1,0.3,0.5,0.7,0.9}\{0.1,0.3,0.5,0.7,0.9\}. Empirically, the size of the word window is set to 3. For DAML, the dimensionality of latent feature vectors is chosen from {8,16,32,64}\{8,16,32,64\}, and the sliding window size is set to 3. The learning rate is selected from {0.00001,0.0001,0.001,0.002,0.01}\{0.00001,0.0001,0.001,0.002,0.01\}. The dropout ratio is selected from {0.1,0.3,0.5,0.7,0.9}\{0.1,0.3,0.5,0.7,0.9\}.

For all evaluated methods, we employ early stopping strategy for model training, and the optimal hyper-parameters are chosen based on the performances on the validation data.

4.2 Performance Comparison

Table II summarizes the results of different recommendation methods. We can make the following observations.

  • The review-based recommendation methods significantly outperforms PMF, which demonstrates the importance of review data for improving recommendation accuracy.

  • The neural methods using CNNs and attention mechanism (e.g., DeepCoNN and NARRE) achieve better performances than CDL. The potential reason is that CDL employs bag-of-words model to process the review data, thus ignores the word order and local context information.

  • The review-based methods based on attention mechanism (e.g., NARRE, NRPA, RMG) outperform DeepCoNN, by learning the importance of words or reviews for better understanding of users’ rating behaviours. DAML usually performs best among baseline methods. It utilizes mutual attention of the CNN to study the correlation between the user reviews and item reviews, thus can capture more information about the user preferences and item properties.

  • RGNN consistently achieves the best performances on all datasets. The improvements over baselines are statistically significant with p<0.05p<0.05 using student t-test. In RGNN, the user/item reviews are described by a directed review graph with preserving the word orders, which can capture the global and non-consecutive topology and dependency information between words in reviews. Moreover, the PGP operator extracts hierarchical and coarser-grained representations of reviews while considering the user/item-specific properties, thus can further help improve the recommendation performance.

Refer to caption
(a) Music
Refer to caption
(b) Office
Refer to caption
(c) Toys
Figure 4: Performances of RGNN-T, RGNN-P, RGNN-S, RGNN-PL, and RGNN on Music, Office, and Toys datasets.

Moreover, Table III summarizes the performances of RGNN with respect to different settings of LL. On Music, Office, Toys datasets, the best performances of RGNN are achieved by setting LL to 2. On Games, Beauty, and Yelp datasets, RGNN achieves the best performances, when LL is set to 3, 4, and 3, respectively. This demonstrates the effectiveness of the hierarchical graph representation learning.

4.3 Ablation Study

To investigate the importance of each component of RGNN, we firstly study the contributions of the review graph and the factorization machine layer. We consider the following variants of RGNN for experiments:

  • RGNN-RG: In this variant, we remove both the user and item review graphs. We use the max pooling of the embeddings of the keywords extracted from 𝒮u\mathcal{S}_{u} as the review representation of the user uu, which is denoted by 𝐫u\mathbf{r}_{u}. Then, in Eq. (10), we concatenate 𝐦u\mathbf{m}_{u} with 𝐫u\mathbf{r}_{u} to form the user representation 𝐩u\mathbf{p}_{u}. Similar operation is performed on the reviews of the item ii. Note that FM layer is also used for rating prediction in this variant.

  • RGNN-FM: For this variant, we remove the FM layer and replace the rating prediction function in Eq. (11) by the product of 𝐩u\mathbf{p}_{u} and 𝐪i\mathbf{q}_{i} as y^ui=𝐩u𝐪i\hat{y}_{ui}=\mathbf{p}_{u}\mathbf{q}_{i}^{\top}.

Table IV summarizes the performances achieved by DAML, RGNN-RG, RGNN-FM, and RGNN on the Music, Office, and Toys datasets. From Table IV, we can note that RGNN outperforms RGNN-RG by 5.01% and 1.48% on Music and Toys dataset. Compared with the best baseline method DAML that employs FM layer for prediction, RGNN-FM outperforms DAML on the Office and Toys datasets, achieves comparable results with DAML on the Music dataset. These two observations demonstrate the effectiveness of the proposed review graph based representation learning strategy in capturing the users’ preferences on items. Moreover, RGNN outperforms RGNN-FM on Music and Toys datasets. This indicates the FM layer can help improve the rating prediction accuracy. As shown in Table IV, RGNN-FM achieves better results than RGNN-RG on three datasets. This demonstrates that the main improvements achieved by RGNN are caused by the proposed hierarchical review graph representation learning strategy.

Refer to caption
(a) Music
Refer to caption
(b) Office
Refer to caption
(c) Toys
Figure 5: The performances of RGNN with different settings of edge types on Music, Office, and Toys datasets.
Refer to caption
(a) d0d_{0}, d1d_{1}, and d2d_{2}
Refer to caption
(b) α\alpha
Refer to caption
(c) ω\omega
Figure 6: The performances of RGNN with respect to different settings of d0d_{0}, d1d_{1}, d2d_{2}, α\alpha, and ω\omega on Music dataset.
TABLE IV: Performances of DAML, RGNN-RG, RGNN-FM, and RGNN on the Music, Office, and Toys datasets.
Dataset DAML RGNN-RG RGNN-FM RGNN
Music 0.7401 0.7705 0.7411 0.7319
Office 0.7164 0.7132 0.7115 0.7125
Toys 0.7909 0.7903 0.7817 0.7786

Moreover, we perform experiments to analyze the contributions of different components in the hierarchical graph representation learning procedure. Specifically, we evaluate the performances of the following RGNN variants:

  • RGNN-T: RGNN without considering the word orders (i.e., removing the edge type embedding in Eq. (2)).

  • RGNN-P: RGNN without considering personalized properties of users/items (i.e., using the same 𝜽\bm{\theta}^{\ell} for all users/items in Eq. (5)).

  • RGNN-S: RGNN without considering the feature diversity of a node’s neighborhood in estimating the node importance (i.e., removing the second part in Eq. (5)).

  • RGNN-PL: RGNN without considering the graph pooling (i.e., removing the PGP operators in Figure 3).

The experimental results on Music, Office, and Toys datasets are summarized in Figure 4. As shown in Figure 4, RGNN outperforms RGNN-T on all three datasets. This indicates that the word order is important in extracting semantic information from reviews for recommendation. Moreover, RGNN achieves better performances than RGNN-P, RGNN-S and RGNN-PL in terms of MSE. This result demonstrates that the proposed PGP operator can help improve the recommendation accuracy via capturing more informative review words for a user/item. When estimating the importance of a node (i.e., word), the proposed PGP operator considers the user/item-specific properties, as well as the feature diversity of a node’s neighborhood in the review graph. From Figure 4, we can also note that GNN-PL achieves better MSE than RGNN-P on Office and Toys datasets. RGNN-P adopts the same pooling parameter for all users/items. It tends to select a set of common words to form the review representations for all the users/items. The specific properties of a user/item in the review data may not be exploited. Therefore, RGNN-P may perform poorer than RGNN-PL that employs max pooling on the embeddings of all nodes in the review graph of each individual user/item.

Refer to caption
Figure 7: Highlighted words according to importance scores in the review documents of a user-item pair. The words with dark yellow background have larger importance scores than the words with yellow background. The words with red background are manually selected to show the user’s opinions on the item.

4.4 Impacts of Edge Types

In addition, we also conduct experiments to analyze the impacts of different edge types. Specifically, we study the performances of RGNN with the following settings:

  • RGNN-f: RGNN without considering the forward edge type efe_{f}.

  • RGNN-b: RGNN without considering the backward edge type ebe_{b}.

  • RGNN-s: RGNN without considering the self-loop edge type.

Figure 5 summarizes the MSE values achieved by RGNN with different settings of edge types on Music, Office, and Toys datasets. As shown in Figure 5, RGNN-b outperforms RGNN-f on all three datasets. This indicates that the forward type of edges are more important than the backward type of edges in capturing users’ rating behaviors for recommendation. Moreover, RGNN-s achieves the worst MSE values on Music and Toys datasets. This indicates the self-loop type of edges play an important role in the graph representation learning procedure. This observation is consistent with the definition of the attention mechanism in Eq. (3). By removing the self-loop type of edges, for each node xhx_{h} in the graph at the \ell-th layer, it is not included in its first-hop neighbors 𝒩h\mathcal{N}_{h}^{\ell} in the graph. According to Eq. (3), the new embedding of xhx_{h} (i.e., 𝐱¯h\bar{\mathbf{x}}_{h}^{\ell}) is updated by only considering the input embeddings of other neighboring nodes at the \ell-th layer, without considering the input embedding of xhx_{h} at the \ell-th layer. In addition, RGNN achieves the best results on Music and Office datasets. This indicates that the combination of all three types of edges usually provides a better description about the review data, thus achieves more accurate rating prediction results.

4.5 Parameter Sensitivity Study

In this section, we perform experiments to evaluate the sensitivity to hyper-parameters for RGNN. The dimensionality of the semantic space d0d_{0}, the dimensionality of the rating space d1d_{1}, and the dimensionality of the hidden space of the graph representation learning layers d2d_{2}, are varied in {4,8,16,32,48,64,128}\{4,8,16,32,48,64,128\}. The graph pooling ratio α\alpha is adjusted in {0.1,0.3,0.5,0.7,0.9}\{0.1,0.3,0.5,0.7,0.9\}. The size of the word window ω\omega is chosen from {2, 3, 4, 5, 6, 7, 8, 9, 10}. The results on Music dataset are summarized in Figure 6. We can note that the best MSE can be achieved by setting d0d_{0}, d1d_{1} to 16, and d2d_{2} to 8. Larger dimensionality of the latent semantic space and rating space cause more computation cost, however may not help improve the rating prediction accuracy. For the pooling ratio, we can notice that the best setting for α\alpha on Music dataset is 0.5. Moreover, RGNN usually achieves better performances by setting α<0.5\alpha<0.5, compared by setting α>0.5\alpha>0.5. One potential reason is that when α>0.5\alpha>0.5 more nodes would be kept by the pooling operation, thus the “noise nodes” are more likely to be involved in learning the hierarchical review graph representations. In addition, as shown in Figure 6 (c), RGNN achieves the best MSE value when ω\omega is set to 5 on Music dataset. Better performances can be usually achieved when ω\omega is set in the range between 3 and 6.

4.6 Case Study

In addition, we also conduct a case study to further explore whether RGNN can select informative words in the graph pooling layer. We sample a user-item pair in test data and extract the review documents of the user and item. Figure 7 visualizes the heat maps of these review documents. We highlight the words extracted from the first PGP layer of RGNN. The words with dark yellow background have larger importance scores than the words with yellow background (refer to Eq. (5)). As shown in Figure 7, the words “playing”, “sound”, “money”, and “long” are highlighted in user reviews. These words indicate that this user may be more interested in the quality and price of an item. Moreover, in the item reviews, the words “high”, “quality”, “solid”, “sound”, and “cheap” which describe the item properties are selected in the first pooling layer. In the review given by the user on the item, from the words “price”, “quality”, “natural”, and “sound”, we can argue that the user’s high rating on this item is indeed, because this item meets the concerns of the user. Hence, this case again demonstrates that the PGP layer can select informative words from the review graphs, thus can help improve the recommendation accuracy.

5 Conclusion and Future Work

In this work, we propose a novel review-based recommendation model, namely Review Graph Neural Network (RGNN), which incorporates the advantages of graph attention networks and hierarchical graph pooling in understanding users’ review behaviors for recommendation. In RGNN, the reviews of each user/item are described by a directed graph, which simultaneously considers the co-occurrence of words and the word orders. RGNN employs a type-aware graph attention network and a personalized graph pooling operator to learn hierarchical semantic representations for users and items from the review data. A factorization machine layer is then developed to predict the user’s rating on an item, considering the interactions between the user and item semantic features. The experimental results on two real-world datasets show that RGNN consistently outperforms state-of-the-art review-based recommendation methods, in terms of MSE. For future work, we would like to study how to maintain the connectivity of the review graph after the graph pooling operation. Moreover, we are also interested in extending RGNN to solve the top-N item recommendation problem.

References

  • [1] Y. Shi, M. Larson, and A. Hanjalic, “Collaborative filtering beyond the user-item matrix: A survey of the state of the art and future challenges,” CSUR, vol. 47, no. 1, p. 3, 2014.
  • [2] S. Zhang, L. Yao, A. Sun, and Y. Tay, “Deep learning based recommender system: A survey and new perspectives,” CSUR, vol. 52, no. 1, p. 5, 2019.
  • [3] L. Hu, A. Sun, and Y. Liu, “Your neighbors affect your ratings: on geographical neighborhood influence to rating prediction,” in Proceedings of the 37th international ACM SIGIR conference on Research & development in information retrieval, 2014, pp. 345–354.
  • [4] X. Liu, Y. Liu, K. Aberer, and C. Miao, “Personalized point-of-interest recommendation by mining users’ preference transition,” in Proceedings of the 22nd ACM international conference on Information & Knowledge Management, 2013, pp. 733–738.
  • [5] Q. Wu, Y. Liu, C. Miao, B. Zhao, Y. Zhao, and L. Guan, “Pd-gan: adversarial learning for personalized diversity-promoting recommendation,” in IJCAI’19.   AAAI Press, 2019, pp. 3870–3876.
  • [6] Z. Sun, J. Yang, J. Zhang, A. Bozzon, L.-K. Huang, and C. Xu, “Recurrent knowledge graph embedding for effective recommendation,” in RecSys’18.   ACM, 2018, pp. 297–305.
  • [7] L. Chen, G. Chen, and F. Wang, “Recommender systems based on user reviews: the state of the art,” User Modeling and User-Adapted Interaction, vol. 25, no. 2, pp. 99–154, 2015.
  • [8] D. M. Blei, A. Y. Ng, and M. I. Jordan, “Latent dirichlet allocation,” JMLR, vol. 3, no. Jan, pp. 993–1022, 2003.
  • [9] T. Mikolov, K. Chen, G. Corrado, and J. Dean, “Efficient estimation of word representations in vector space,” arXiv preprint arXiv:1301.3781, 2013.
  • [10] J. McAuley and J. Leskovec, “Hidden factors and hidden topics: understanding rating dimensions with review text,” in RecSys’13, 2013, pp. 165–172.
  • [11] Y. Bao, H. Fang, and J. Zhang, “Topicmf: Simultaneously exploiting ratings and reviews for recommendation,” in AAAI’14, 2014.
  • [12] G. Ling, M. R. Lyu, and I. King, “Ratings meet reviews, a combined approach to recommend,” in RecSys’14, 2014, pp. 105–112.
  • [13] Y. Tan, M. Zhang, Y. Liu, and S. Ma, “Rating-boosted latent topics: Understanding users and items with ratings and reviews.” in IJCAI’16, vol. 16, 2016, pp. 2640–2646.
  • [14] W. Zhang, Q. Yuan, J. Han, and J. Wang, “Collaborative multi-level embedding learning from reviews for rating prediction.” in IJCAI’16, 2016, pp. 2986–2992.
  • [15] H. Wang, N. Wang, and D.-Y. Yeung, “Collaborative deep learning for recommender systems,” in KDD’15.   ACM, 2015, pp. 1235–1244.
  • [16] L. Zheng, V. Noroozi, and P. S. Yu, “Joint deep modeling of users and items using reviews for recommendation,” in WSDM’17.   ACM, 2017, pp. 425–434.
  • [17] C. Chen, M. Zhang, Y. Liu, and S. Ma, “Neural attentional rating regression with review-level explanations,” in WWW’18, 2018, pp. 1583–1592.
  • [18] D. Liu, J. Li, B. Du, J. Chang, and R. Gao, “Daml: Dual attention mutual learning between ratings and reviews for item recommendation,” in KDD’19.   ACM, 2019, pp. 344–352.
  • [19] H. Liu, F. Wu, W. Wang, X. Wang, P. Jiao, C. Wu, and X. Xie, “Nrpa: Neural recommendation with personalized attention,” SIGIR’19, 2019.
  • [20] K. Bauman, B. Liu, and A. Tuzhilin, “Aspect based recommendations: Recommending items with the most valuable aspects based on user reviews,” in Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2017, pp. 717–725.
  • [21] J. Y. Chin, K. Zhao, S. Joty, and G. Cong, “Anr: Aspect-based neural recommender,” in Proceedings of the 27th ACM International Conference on Information and Knowledge Management, 2018, pp. 147–156.
  • [22] L. Wu, C. Quan, C. Li, Q. Wang, B. Zheng, and X. Luo, “A context-aware user-item representation learning for item recommendation,” TOIS, vol. 37, no. 2, p. 22, 2019.
  • [23] Y. Lu, R. Dong, and B. Smyth, “Coevolutionary recommendation model: Mutual learning between ratings and reviews,” in WWW’18, 2018, pp. 773–782.
  • [24] C. Wu, F. Wu, T. Qi, S. Ge, Y. Huang, and X. Xie, “Reviews meet graphs: Enhancing user and item representations for recommendation with hierarchical attentive graph neural network,” in EMNLP-IJCNLP 2019, 2019, pp. 4886–4895.
  • [25] Z. Wu, S. Pan, F. Chen, G. Long, C. Zhang, and S. Y. Philip, “A comprehensive survey on graph neural networks,” IEEE Transactions on Neural Networks and Learning Systems, 2020.
  • [26] F. Scarselli, M. Gori, A. C. Tsoi, M. Hagenbuchner, and G. Monfardini, “The graph neural network model,” IEEE Transactions on Neural Networks, vol. 20, no. 1, pp. 61–80, 2008.
  • [27] C. Gallicchio and A. Micheli, “Graph echo state networks,” in The 2010 International Joint Conference on Neural Networks (IJCNN).   IEEE, 2010, pp. 1–8.
  • [28] Y. Li, D. Tarlow, M. Brockschmidt, and R. Zemel, “Gated graph sequence neural networks,” in ICLR, 2015.
  • [29] J. Bruna, W. Zaremba, A. Szlam, and Y. Lecun, “Spectral networks and locally connected networks on graphs,” in ICLR, 2014.
  • [30] M. Defferrard, X. Bresson, and P. Vandergheynst, “Convolutional neural networks on graphs with fast localized spectral filtering,” in NIPS, 2016, pp. 3844–3852.
  • [31] T. N. Kipf and M. Welling, “Semi-supervised classification with graph convolutional networks,” arXiv preprint arXiv:1609.02907, 2016.
  • [32] A. Micheli, “Neural network for graphs: a contextual constructive approach,” IEEE Transactions on Neural Networks, vol. 20, no. 3, pp. 498–511, 2009.
  • [33] J. Atwood and D. Towsley, “Diffusion-convolutional neural networks,” in Advances in Neural Information Processing Systems 29, 2016, pp. 1993–2001.
  • [34] W. Hamilton, Z. Ying, and J. Leskovec, “Inductive representation learning on large graphs,” in Advances in Neural Information Processing Systems 30, 2017, pp. 1024–1034.
  • [35] P. Veličković, G. Cucurull, A. Casanova, A. Romero, P. Lio, and Y. Bengio, “Graph attention networks,” arXiv preprint arXiv:1710.10903, 2017.
  • [36] J. Zhang, X. Shi, J. Xie, H. Ma, I. King, and D. Y. Yeung, “Gaan: Gated attention networks for learning on large and spatiotemporal graphs,” in UAI, 2018.
  • [37] J. L. Keyulu Xu, Weihua Hu and S. Jegelka, “How powerful are graph neural networks?” in ICLR, 2019.
  • [38] M. Defferrard, X. Bresson, and P. Vandergheynst, “Convolutional neural networks on graphs with fast localized spectral filtering,” in Advances in neural information processing systems, 2016, pp. 3844–3852.
  • [39] M. Fey, J. Eric Lenssen, F. Weichert, and H. Müller, “Splinecnn: Fast geometric deep learning with continuous b-spline kernels,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 869–877.
  • [40] F. M. Bianchi, D. Grattarola, and C. Alippi, “Mincut pooling in graph neural networks,” arXiv preprint arXiv:1907.00481, 2019.
  • [41] Y. Ma, S. Wang, C. C. Aggarwal, and J. Tang, “Graph convolutional networks with eigenpooling,” arXiv preprint arXiv:1904.13107, 2019.
  • [42] O. Vinyals, S. Bengio, and M. Kudlur, “Order matters: Sequence to sequence for sets,” arXiv preprint arXiv:1511.06391, 2015.
  • [43] M. Zhang, Z. Cui, M. Neumann, and Y. Chen, “An end-to-end deep learning architecture for graph classification,” in Thirty-Second AAAI Conference on Artificial Intelligence, 2018.
  • [44] Z. Ying, J. You, C. Morris, X. Ren, W. Hamilton, and J. Leskovec, “Hierarchical graph representation learning with differentiable pooling,” in Advances in Neural Information Processing Systems, 2018, pp. 4800–4810.
  • [45] H. Gao and S. Ji, “Graph u-nets,” arXiv preprint arXiv:1905.05178, 2019.
  • [46] H. Gao, Y. Chen, and S. Ji, “Learning graph pooling and hybrid convolutional operations for text representations,” in WWW’19.   ACM, 2019, pp. 2743–2749.
  • [47] J. Lee, I. Lee, and J. Kang, “Self-attention graph pooling,” arXiv preprint arXiv:1904.08082, 2019.
  • [48] F. Diehl, “Edge contraction pooling for graph neural networks,” arXiv preprint arXiv:1905.10990, 2019.
  • [49] E. Ranjan, S. Sanyal, and P. P. Talukdar, “Asap: Adaptive structure aware pooling for learning hierarchical graph representations,” arXiv preprint arXiv:1911.07979, 2019.
  • [50] Z. Zhang, J. Bu, M. Ester, J. Zhang, C. Yao, Z. Yu, and C. Wang, “Hierarchical graph pooling with structure learning,” arXiv preprint arXiv:1911.05954, 2019.
  • [51] S. Rendle, “Factorization machines,” in ICDM’10, 2010, pp. 995–1000.
  • [52] F. D. Malliaros and M. Vazirgiannis, “Graph-based text representations: Boosting text mining, nlp and information retrieval with graphs,” in EMNLP’17: Tutorial Abstracts, 2017.
  • [53] R. Mihalcea and P. Tarau, “TextRank: Bringing order into text,” in EMNLP’04, Jul. 2004, pp. 404–411.
  • [54] D. Shen, G. Wang, W. Wang, M. R. Min, Q. Su, Y. Zhang, C. Li, R. Henao, and L. Carin, “Baseline needs more love: On simple word-embedding-based models and associated pooling mechanisms,” in ACL’18, 2018, pp. 440–450.
  • [55] R. He and J. McAuley, “Ups and downs: Modeling the visual evolution of fashion trends with one-class collaborative filtering,” in WWW’16, 2016, pp. 507–517.
  • [56] A. Mnih and R. R. Salakhutdinov, “Probabilistic matrix factorization,” in NIPS’08, 2008, pp. 1257–1264.
  • [57] Y. Koren, “Factorization meets the neighborhood: a multifaceted collaborative filtering model,” in Proceedings of the 14th ACM SIGKDD international conference on Knowledge discovery and data mining, 2008, pp. 426–434.
  • [58] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.