This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Comparative Explanations of Recommendations

Aobo Yang1, Nan Wang1, Renqin Cai1, Hongbo Deng2, Hongning Wang1 1University of Virginia, Charlottesville USA 2Alibaba Group, Hangzhou China ay6gv, nw6a, rc7ne@virginia.edu, dhb167148@alibaba-inc.com, hw5x@virginia.edu
(2022)
Abstract.

As recommendation is essentially a comparative (or ranking) process, a good explanation should illustrate to users why an item is believed to be better than another, i.e., comparative explanations about the recommended items. Ideally, after reading the explanations, a user should reach the same ranking of items as the system’s. Unfortunately, little research attention has yet been paid on such comparative explanations.

In this work, we develop an extract-and-refine architecture to explain the relative comparisons among a set of ranked items from a recommender system. For each recommended item, we first extract one sentence from its associated reviews that best suits the desired comparison against a set of reference items. Then this extracted sentence is further articulated with respect to the target user through a generative model to better explain why the item is recommended. We design a new explanation quality metric based on BLEU to guide the end-to-end training of the extraction and refinement components, which avoids generation of generic content. Extensive offline evaluations on two large recommendation benchmark datasets and serious user studies against an array of state-of-the-art explainable recommendation algorithms demonstrate the necessity of comparative explanations and the effectiveness of our solution.

explainable recommendation, comparative explanation, text generation, extract-and-refine
journalyear: 2022copyright: rightsretainedconference: Proceedings of the ACM Web Conference 2022; April 25–29, 2022; Virtual Event, Lyon, Francebooktitle: Proceedings of the ACM Web Conference 2022 (WWW ’22), April 25–29, 2022, Virtual Event, Lyon, Francedoi: 10.1145/3485447.3512031isbn: 978-1-4503-9096-5/22/04ccs: Information systems Recommender systemsccs: Computing methodologies Natural language generation

1. Introduction

Modern recommender systems fundamentally shape our everyday life (Koren et al., 2009; He et al., 2017; Sarwar et al., 2001; Rendle, 2010; Aggarwal et al., 2016; Wu et al., 2020; Cai et al., 2021). As a result, how to explain the algorithm-made recommendations becomes crucial in building users’ trust in the systems (Zhang and Chen, 2020). Previous research shows that explanations, which illustrate how the recommendations are generated (Ribeiro et al., 2016; Lundberg and Lee, 2017) or why the users should pay attention to the recommendations (Wang et al., 2018b; Sun et al., 2020; Yang et al., 2021), can notably strengthen user engagement with the system and better assist them in making informed decisions (Bilgic and Mooney, 2005; Herlocker et al., 2000; Sinha and Swearingen, 2002).

When being presented with a list of recommendations, typically sorted in a descending order, a user needs to make a choice. In other words, the provided explanations should help users compare the recommended items. Figure 1 illustrates the necessity of comparative explanations. By reading the explanations for the hotels recommended in the figure, one can easily tell why the system ranks them in such an order. But if the system provided the explanation in the dashed box for Hotel C, it would confuse the users about the ranking, e.g., Hotel C becomes arguably comparable to top ranked Hotel A; but it was ranked at the bottom of the list. This unfortunately hurts users’ trust in all three recommended hotels.

Refer to caption
Figure 1. An illustration about the necessity of comparative explanations. The recommended Hotel A, B, C are listed in a descending order, with the provided explanations to justify the ranking. But if we replace Hotel C’s explanation with the one in the dash box, users may no longer perceive the ranking of all three hotels.
\Description

Example of comparative explanations

Existing explainable recommendation solutions are not optimized to help users make such comparative decisions for two major reasons. First, the explanation of a recommended item is often independently generated without considering other items in the recommendation list. As shown in Figure 1, one single low-quality generation (the one in the dashed box) might hamper a user’s understanding over the entire list of recommendations. Second, the popularly adopted neural text generation techniques are known to be flawed of its generic content output (Holtzman et al., 2019; Welleck et al., 2019). Particularly, techniques like maximum likelihood training and sequence greedy decoding lead to short and repetitive sentences composed of globally frequent words (Weston et al., 2018). Such generic content cannot fulfill the need to differentiate the recommended items. Consider the example shown in Figure 1 again, “the hotel is good” is a very generic explanation and thus not informative. Its vague description (e.g., the word “good”) and lacks of specificity (e.g., the word “hotel”) make it applicable to many hotels, such that users can hardly tell the relative comparison of the recommended items from such explanations.

In this work, we tackle the problem of comparative explanation generation to help users understand the comparisons between the recommended items. We focus on explaining how one item is compared with another; then by using a commonly shared set of items as references (e.g., items the user has reviewed before), the comparisons among the recommended items emerge. For example, if the explanations suggest item A is better than item B and item C is worse than item B, the comparison between A and C is apparent after reading the associated explanations. Our solution is designed to generically work on top of other existing recommender systems. We do not have any assumptions about how the recommendation algorithm ranks items (e.g., collaborative filtering (Sarwar et al., 2001) or content-based (Balabanović and Shoham, 1997)), but only require it to provide a ranking score for each item to our model (i.e., ordinal ranking) which reflects a user’s preference over the recommended item. This makes our solution readily applicable to explain plenty of effective recommendation algorithms deployed in practice.

We design an extract-and-refine text generation architecture (Weston et al., 2018; Guu et al., 2018) to explain the ranked items one at a time to the user, conditioned on their recommendation scores and associated reviews. We refer to the item to be explained in the ranked list as the target item, and user we are explaining to as the target user. First, the model extracts one sentence from the existing review sentences about the target item as a prototype, with a goal to maximize the likelihood of fitting the comparisons against the reviews written by the target user for other reference items. Then we refine the extracted prototype through a generative model to further polish the content for the target user. In this two stage procedure, the extraction module exploits the content already provided about the target item to ensure the relevance of generated explanations (e.g., avoid mentioning features that do not exist in the target item); and the refinement module further improve the explanation (e.g., informativeness and diversity of content) beyond the limitation of the existing content. We design a new explanation quality metric based on BLEU to guide the end-to-end training of the two modules, with a particular focus to penalize short and generic content in generated explanations.

We compared the proposed solution with a rich set of state-of-the-art baselines for explanation generation on two large-scale recommendation datasets. Besides, we also conducted extensive user studies to have the generated explanations evaluated by real users. Positive results obtained on both offline and online experiments suggested the effectiveness of comparative explanations in assisting users to better understand the recommendations and make more informed choices.

2. Related Work

Most explainable recommendation solutions exploit user reviews as the source of training data. They either directly extract from reviews or synthesize content to mimic the reviews. Extraction-based solutions directly select representative text snippets from the target item’s existing reviews. For example, NARRE (Chen et al., 2018) selects the most attentive reviews as the explanation, based on the attention that is originally learned to enrich the user and item representations for recommendation. CARP (Li et al., 2019) uses the capsule network for the same purpose. Wang et al. (2018a) adopt reinforcement learning to extract the most relevant review text that matches a given recommender system’s rating prediction. Xian et al. (2021) extract attributes from reviews to explain a set of items based on users’ preferences. However, as such solutions are restricted to an item’s existing reviews, their effectiveness is subject to the availability and quality of existing content. For items with limited exposure, e.g., a new item, these solutions can hardly provide any informative explanations.

Generation-based solutions synthesize textual explanations that are not limited to existing reviews. One branch focuses on predicting important aspects of an item (such as item features) from its associated reviews as explanations (Wang et al., 2018b; Tao et al., 2019; He et al., 2015; Ai et al., 2018; Cai et al., 2017). For instance, MTER (Wang et al., 2018b) and FacT (Tao et al., 2019) predict item features that are most important for a user to justify the recommendation. They rely on predefined text templates to deliver the predicted features. The other branch applies neural text generation techniques to synthesize natural language sentences. In particular, NRT (Li et al., 2017) models item recommendation and explanation generation in a shared user and item embedding space. It uses its predicted recommendation ratings as part of the initial state for explanation generation. MRG (Truong and Lauw, 2019) integrates multiple modalities from user reviews, including ratings, text, and associated images, for multi-task explanation modeling.

Our work is closely related to two recent studies, DualPC (Sun et al., 2020) and SAER (Yang et al., 2021), which focus on strengthening the relation between recommendations and explanations. Specifically, DualPC introduces duality regularization based on the joint probability of explanations and recommendations to improve the correlation between recommendations and generated explanations. SAER introduces the idea of sentiment alignment in explanation generation. However, both of them operate in a pointwise fashion, i.e., independent explanation generation across items. Our solution focuses on explaining the comparisons between items. We should also emphasize our solution is to explain the comparison among a set of recommended items, rather than to find comparable items (McAuley et al., 2015; Chen et al., 2020).

There are also solutions exploiting other types of information for explainable recommendation, such as item-item relation (Chen et al., 2021), knowledge graph (Xian et al., 2019) and social network (Ji and Shen, 2016). But they are clearly beyond the scope of this work.

3. Comparative Explanation Generation

Item recommendation in essence is a ranking problem: estimate a recommendation score for each item under a given user and rank the items accordingly, such that the utility of the recommendations can be maximized (Rendle et al., 2012; Karatzoglou et al., 2013). Instead of explaining how the recommendation scores are obtained, our work emphasizes on explaining how the comparisons between the ranked items are derived.

To learn the explanation model, we assume an existing corpus of item reviews from the intended application domain (e.g., hotel reviews). Each review is uniquely associated with a user uu and an item cc, and a user-provided rating rcu{r}_{c}^{u} suggesting his/her opinion towards the item. We group the reviews associated with user uu to construct his/her profile Ωu={(x1u,r1u),(x2u,r2u),,(xmu,rmu)}\Omega_{u}=\{(x^{u}_{1},r^{u}_{1}),(x^{u}_{2},r^{u}_{2}),...,(x^{u}_{m},r^{u}_{m})\}, where xiux^{u}_{i} is the ii-th review sentence extracted from user uu’s reviews and riur^{u}_{i} is the corresponding opinion rating. riur^{u}_{i} can be easily obtained when the detailed aspect ratings are available (Wang et al., 2010); otherwise off-the-shelf sentiment analysis methods can be used for the purpose (interested users can refer to (Wang et al., 2018b; Zhang et al., 2014) for more details). As regards cold-start for users without reviews, generic profiles can be used instead which sample reviews from similar users clustered by other non-review-related features, such as rating history. We create the item profile as Ψc={x1c,x2c,,xnc}\Psi_{c}=\{x^{c}_{1},x^{c}_{2},...,x^{c}_{n}\}, where xjcx^{c}_{j} is the jj-th review sentence extracted from item cc’s existing reviews. Unlike the user profile, the item profile does not include ratings. This is because the ratings from different users are not directly comparable, as individuals understand or use the numerical ratings differently. Our solution is agnostic to the number of entries in user profile Ωu\Omega_{u} and item profile Ψc\Psi_{c} in each user and item.

We impose a generative process for a tuple (x,rcu)(x,r^{u}_{c}) from user uu about item cc conditioned on Ψc\Psi_{c} and Ωu\Omega_{u}. We assume when user uu is reviewing item cc, he/she will first select an existing sentence from Ψc\Psi_{c} that is mostly related to the aspect he/she wants to cover about the item. Intuitively, this can be understood as the user will first browse existing reviews of the item to understand how the other users evaluated this item. Then he/she will rewrite this selected sentence to reflect his/her intended opinion and own writing style. This can be considered as a set to sequence generation problem. For our purpose of explanation generation, we only concern the generation of opinionated text xx. Hence, we take opinion rating rcur^{u}_{c} as input, which leads us to the following formulation,

(1) P(x|u,c,rcu)=xjcΨcPref(x|xjc,rcu,Ωu)Pext(xjc|rcu,Ωu)P(x|u,c,r^{u}_{c})=\sum_{x^{c}_{j}\in\Psi_{c}}P_{ref}(x|x^{c}_{j},r^{u}_{c},\Omega_{u})P_{ext}(x^{c}_{j}|r^{u}_{c},\Omega_{u})

where Pext(xjc|rcu,Ωu)P_{ext}(x^{c}_{j}|r^{u}_{c},\Omega_{u}) specifies the probability that xjcx^{c}_{j} from item profile Ψc\Psi_{c} will be selected by user uu, and Pref(x|xjc,rcu,Ωu)P_{ref}(x|x^{c}_{j},r^{u}_{c},\Omega_{u}) specifies the probability that user uu will rewrite xjcx^{c}_{j} into xx. We name the resulting model Comparative Explainer, or CompExp in short.

In Eq (1), Pext(xjc|rcu,Ωu)P_{ext}(x^{c}_{j}|r^{u}_{c},\Omega_{u}) is essential to capture the comparative textual patterns embedded in user uu’s historical opinionated text content. To understand this, we can simply rewrite its condition part: define Δriu=rcuriu\Delta r^{u}_{i}=r^{u}_{c}-r^{u}_{i}, we have (rcu,Ωu)={(xiu,Δriu)}i=1m(r^{u}_{c},\Omega_{u})=\{(x^{u}_{i},\Delta r^{u}_{i})\}^{m}_{i=1}; hence, Pext(xjc|rcu,Ωu)P_{ext}(x^{c}_{j}|r^{u}_{c},\Omega_{u}) characterizes whether the sentence xjcx^{c}_{j} about item cc is qualified to characterize the desired opinion difference conditioned on user uu’s historical content Ωu\Omega_{u} and target rating rcur^{u}_{c}. For example, a negative Δriu\Delta r^{u}_{i} suggests the opinion conveyed in xjcx^{c}_{j} is expected to be less positive than that in xiux^{u}_{i}. On a similar note, Pref(x|xjc,rcu,Ωu)P_{ref}(x|x^{c}_{j},r^{u}_{c},\Omega_{u}) quantifies if xx is a good rewriting of xjcx^{c}_{j} to satisfy the desired opinion rating rcur^{u}_{c} for item cc by user uu.

One can parameterize Pext(xjc|rcu,Ωu)P_{ext}(x^{c}_{j}|r^{u}_{c},\Omega_{u}) and Pref(x|xjc,rcu,Ωu)P_{ref}(x|x^{c}_{j},r^{u}_{c},\Omega_{u}) and estimate the corresponding parameters based on the maximum likelihood principle over observations in Ωu\Omega_{u}. However, data likelihood alone is insufficient to generate high-quality explanations, as we should also emphasize on fluency, brevity, and diversity of the generated explanations. To realize this generalized objective, assume a metric π(x|u,c)\pi(x|u,c) that measures the quality of generated explanation xx for user uu about item cc, the training objective of CompExp is set to maximize the expected quality of its generated explanations under π(x|u,c)\pi(x|u,c),

(2) J=𝔼xP(x|u,c,rcu)[π(x|u,c)]J=\mathbb{E}_{x\sim P(x|u,c,r^{u}_{c})}[\pi(x|u,c)]

In this work, we present a customized BLUE score specifically for the comparative explanation generation problem to penalize short and generic content.

Next, we dive into the detailed design of CompExp in Section 3.1, then present our metric π(x|u,c)\pi(x|u,c) for parameter estimation in Section 3.2 and 3.3, and finally illustrate how to estimate each component in CompExp end-to-end in Section 3.4.

3.1. Extract-and-Refine Architecture

Our proposed model architecture for CompExp is shown in Figure 2, which in a nutshell is a fully connected hierarchical neural network. The explanations for a user item pair (u,c)(u,c) is generated via an extract-and-refine process, formally described in Eq (1). Comparing to existing pure generation-based explanation methods (Li et al., 2017; Sun et al., 2020; Yang et al., 2021), one added benefit of our solution is to ensure faithfulness of the generated explanations: it avoids mentioning attributes that are not relevant to the target item. To address the limitations in directly using existing content, e.g., unaligned content style or sentiment polarity, the refinement step further rewrites the extracted sentence to make its content better fit for the purpose of comparative explanation, e.g., improve the quality defined by π(x|u,c)\pi(x|u,c).

We refer to Pext(xjc|rcu,Ωu)P_{ext}(x^{c}_{j}|r^{u}_{c},\Omega_{u}) as the extractor and Pref(x|xjc,rcu,Ωu)P_{ref}(x|x^{c}_{j},r^{u}_{c},\Omega_{u}) as the refiner. Next, we will zoom into each component to discuss its design principle and technical details.

Refer to caption
Figure 2. The extract-and-refine model architecture for CompExp. The extractor extracts a candidate sentence from item cc’s profile as a prototype for explanation generation; and the refiner rewrites this sentence to optimize the desired quality metric for comparative explanation.

3.1.1. Extractor.

The extractor’s goal is to select a prototype sentence xjcx^{c}_{j} from item cc’s profile Ψc\Psi_{c} for a given opinion rating rcu{r}^{u}_{c} that best satisfies the comparativeness suggested by the user profile Ωu\Omega_{u}. We refer to xjcΨcx^{c}_{j}\in\Psi_{c} as an extraction candidate and xiuΩux^{u}_{i}\in\Omega_{u} as a reference. The extractor adopts a bidirectional GRU (Chung et al., 2014) as the universal text encoder to convert the extraction candidates and references into continuous embedding vectors. Since the pairwise comparison specified by Δriu\Delta r^{u}_{i} is a scalar, we use a one-hot vector to encode it when the ratings are discrete, otherwise we use a non-linear multi-layer perceptron (MLP) as the rating encoder.

Intuitively, in the one dimensional rating space, we can easily recover the intended sentence’s rating rcu{r}^{u}_{c} from the rating of the reference sentence riu{r}^{u}_{i} and required rating difference Δriu\Delta r^{u}_{i}. As an analogy, we consider the rating difference vector as the transform direction that suggests the ideal comparative explanation in the latent text space from a reference sentence xiux^{u}_{i}, denoted as f(xiu,Δriu)hif(x^{u}_{i},\Delta r^{u}_{i})\to h_{i}. As a result, hih_{i} is the text embedding vector for the ideal comparative explanation. The extractor implements such a transformation using an MLP taking the concatenation of the text embedding and rating difference embedding vectors as input.

Given the desired comparative explanation hih_{i}, the extraction candidates can be evaluated by their similarities towards hih_{i}. This specifies a directional distribution Q(x;hi)Q(x;h_{i}) centered on hih_{i} in the latent text embedding space. Since cosine is a commonly used similarity metric for text embeddings, we formulate Q(x;hi)Q(x;h_{i}) as a von Mises-Fisher distribution (Guu et al., 2018) over all the extraction candidates,

Q(x;hi)fvMF(x;hi,κ)=Cp(κ)eκcos(x,hi)Q(x;h_{i})\propto f_{vMF}(x;h_{i},\kappa)=C_{p}(\kappa)e^{\kappa\cos(x,h_{i})}

where fvMF()f_{vMF}(\cdot) is the probability density function, κ\kappa is the concentration parameter, and Cp(κ)C_{p}(\kappa) is a normalization function about kk. Because each reference sentence xiux^{u}_{i} will suggest a different directional distribution, we extend the von Mises-Fisher distribution to cover multiple centriods and define Pext(xjc|rcu,Ωu)P_{ext}(x^{c}_{j}|r^{u}_{c},\Omega_{u}) as follows,

(3) Pext(xjc|rcu,Ωu)xiuΩufvMF(xjc;f(xiu,Δriu),κ)P_{ext}(x^{c}_{j}|r^{u}_{c},\Omega_{u})\propto\sum_{x^{u}_{i}\in\Omega_{u}}f_{vMF}\Big{(}x^{c}_{j};f(x^{u}_{i},\Delta r^{u}_{i}),\kappa\Big{)}

Intuitively, in Eq (3), each ideal embedding hih_{i} suggests which extraction candidate better fits the comparativeness embedded in Ωu\Omega_{u}. The summation over Ωu\Omega_{u} aggregates each reference sentence’s evaluation on candidate sentence xjcx^{c}_{j}. κ\kappa is kept as a hyper-parameter which shapes the extraction probability distribution: a larger κ\kappa value leads to a skewer distribution. We can use it to control the exploration of the extraction candidates during the policy gradient based model training, which will be introduced in Section 3.4.

3.1.2. Refiner.

The objective of the refiner is to rewrite the extracted prototype to further improve the quality metric π(x|u,c)\pi(x|u,c). As we argued before, a better explanation should be more supportive to the pairwise comparison required by the user profile. Therefore, assuming the refiner successfully turns the prototype xjcx^{c}_{j} into a better framed sentence x^jc\hat{x}^{c}_{j} about the item cc for user uu, then when we give x^jc\hat{x}^{c}_{j} back to the extractor together with xjcx^{c}_{j}, the extractor should prefer the revised version over the original one. Otherwise, we should keep refining x^jc\hat{x}^{c}_{j} until the extractor believes it can no longer be improved. Hence, the refiner needs to find a direction such that Pext(xjc|rcu,Ωu)<Pext(x^jc|rcu,Ωu)P_{ext}(x^{c}_{j}|r^{u}_{c},\Omega_{u})<P_{ext}(\hat{x}^{c}_{j}|r^{u}_{c},\Omega_{u}), which is exactly suggested by the gradient of Pext(xjc|rcu,Ωu)P_{ext}(x^{c}_{j}|r^{u}_{c},\Omega_{u}) with respect to xjcx^{c}_{j}, i.e., the fastest direction for xjcx^{c}_{j} to increase the value of Pext(xjc|rcu,Ωu)P_{ext}(x^{c}_{j}|r^{u}_{c},\Omega_{u}). As a result, our refiner simply pushes the text embedding vector of xjcx^{c}_{j} alone this gradient direction:

zj=xjcPext(xjc|rcu,Ωu)imeκcos(xjc,hi)[hi|xjc||hi|cos(xjc,hi)xjc|xjc|2]\begin{split}{z}_{j}&=\nabla_{x^{c}_{j}}P_{ext}(x^{c}_{j}|r^{u}_{c},\Omega_{u})\\ &\propto\sum^{m}_{i}e^{\kappa\cos(x^{c}_{j},h_{i})}\big{[}\frac{h_{i}}{|x^{c}_{j}||h_{i}|}-\cos(x^{c}_{j},h_{i})\frac{x^{c}_{j}}{|x^{c}_{j}|^{2}}\big{]}\end{split}

Since the refinement step should only polish the extracted prototype instead of dramatically changing it, we normalize the gradient to a unit vector and restrict the step size to one in all cases, i.e., x^jc=xjc+zj/|zj|\hat{x}^{c}_{j}=x^{c}_{j}+z_{j}/|z_{j}|. At last, we include a single-layer GRU with attention (Luong et al., 2015) as the text decoder to convert the refined text vector x^jc\hat{x}^{c}_{j} to the final explanation sentence xx.

Connecting these two modules together, CompExp generates explanations for a ranked list of recommended items one at a time. To understand why the generated explanations carry comparativeness, we can consider the user’s profile Ωu\Omega_{u} as an anchor. Because all the explanations are generated against this anchor, the comparisons among the explanations emerge.

3.2. Explanation Quality Metric

To train CompExp under Eq (2), we need to define the explanation quality metric π(x|u,c)\pi(x|u,c). There is no commonly agreed offline metric for explanation quality in the community yet. And obtaining real user feedback is not feasible for offline model training. Currently, most of explainable recommendation solutions (Sun et al., 2020; Li et al., 2017; Yang et al., 2021) adopt metrics measuring the overlapping content between the generated explanations and user reviews, such as BLEU (Papineni et al., 2002).

However, the BLEU metric, which is initially designed for machine translation, is problematic in explanation evaluation for at least two important reasons. First, it is biased towards shorter sentences. As a precision-based metric, BLEU overcomes the short-length issue by introducing the brevity penalty, which down-scales the precision when the generated length is smaller than its “best match length” (Papineni et al., 2002). The “best match length” design is reasonable in machine translation, because all reference sentences are valid translations covering the information contained in the source language, regardless of their length differences. However, when using review sentences as proxies of explanations, the reference sentences from one review can describe totally different aspects of the same item and vary significantly in length and information contained. Since short-length generation benefits precision (less prone to erroneous word choices), BLEU favors explanations exploiting the short references as the “best match”. As a result, it pushes the models to generate explanations that are generally much shorter than the average sentence length in a review, and hence fails to explain the item in details. Second, though precision-based, BLEU is incapable to differentiate the importance of different words in a reference sentence. Words are valued equally in machine translation, but their impact in explanations varies significantly to users. For example, in Figure 1, the feature and descriptive words such as “swimming pool” and “friendly” help users better understand the target item than a very frequent but generic word, like “hotel” and “good”. BLEU’s indiscrimination to words unavoidably favors the explanations with more generic content due to their higher chance of appearance. We later demonstrate how the BLEU metric led to both short and generic explanations in our experiments.

To design a more appropriate metric to evaluate the explanation quality and better guide our model training, we propose IDF-BLEU, i.e., Inverse Document Frequency (IDF) enhanced BLEU. It introduces three changes on top of BLEU to balance the important factors in explanations: length, content overlapping, and content rarity.

First, to penalize an overly short generation, we replace the “best match length” in the brevity factor with the average length of sentences from all reviews,

BPlen=emin(1lrlx,0)BP_{len}=e^{\min(1-\frac{l_{r}}{l_{x}},0)}

where lrl_{r} and lxl_{x} is the average length of references and the length of the explanation respectively. Second, to differentiate the importance of different words, we introduce IDF to measure the value of n-grams and use it to reweigh the precision in BLEU. We compute the IDF of word gg by the number of sentences where it occurs,

IDF(g)=logSsg+1IDF(g)=log\frac{S}{s_{g}}+1

where SS is the total number of review sentences in the training corpus and sgs_{g} is the number of sentences containing word gg. We approximate the IDF of an n-gram by the largest IDF of its constituent words. Then the clipped n-gram precision in BLEU is modified as

(4) pn=gnxIDF(gn)Countclip(gn)gnxIDF(gn)Count(gn)p_{n}=\frac{\sum_{g^{n}\in x}IDF(g^{n})\cdot Count_{clip}(g^{n})}{\sum_{g^{n}\in x}IDF(g^{n})\cdot Count(g^{n})}

where gng^{n} represents the n-gram and Countclip(gn)Count_{clip}(g^{n}) is the BLEU’s operation to calculate the count of gng^{n} in sentence xx while being clipped by the corresponding maximum count in the references. Through the reweighing, correctly predicting an informative word becomes more rewarding than a generic word. However, it alone cannot evaluate content rarity, since the precision-based metric cannot punish sentences for not including rare words. Therefore, at last, inspired by the length brevity factor in original BLEU, we introduce a similar IDF brevity factor to punish sentences lacking words with high IDF,

BPIDF=emin(1drdx,0)BP_{IDF}=e^{\min(1-\frac{d_{r}}{d_{x}},0)}

where dxd_{x} is the average IDF per word dx=gxIDF(g)/lxd_{x}=\sum_{g\in x}IDF(g)/l_{x} and drd_{r} is corresponding average value in references. Then combining them forms our IDF-BLEU,

(5) IDFBLEU=BPlenBPIDFexp(n=1Nwnlogpn)IDF\mathchar 45\relax BLEU=BP_{len}\cdot BP_{IDF}\cdot\exp\Big{(}\sum^{N}_{n=1}w_{n}\log p_{n}\Big{)}

where wnw_{n} is BLEU’s parameter used as the weight of the n-gram precision. We use the proposed IDF-BLEU as the quality metric π(x|u,c)\pi(x|u,c) for CompExp training.

3.3. Hierarchical Rewards

CompExp is a fully connected neural network which can be trained end-to-end with the gradient derived from Eq (2). However, blind end-to-end training faces the risk that the model violates the purpose of our designed extract-and-refine procedure, as the model has a great degree of freedom to arbitrarily push the prototype xjcx^{c}_{j} in the continuous vector space to optimize Eq (2). For example, it could disregard the extracted prototype and generate totally irrelevant content to the target item cc in the refiner.

To enforce the extract-and-refine workflow, we introduce additional intrinsic reward (Vezhnevets et al., 2017) for each layer respectively to regularize their behaviours. Specifically, as IDF-BLEU is used to measure the explanation quality in Eq (2), we directly use the extracted sentence’s IDF-BLEU to reward the extractor, i.e., introduce πext(xjc|u,c)=IDFBLEU(xjc)\pi_{ext}(x^{c}_{j}|u,c)=IDF\mathchar 45\relax BLEU(x^{c}_{j}). For the refiner, we discourage it in pushing the final generation too far away from the extracted one. Inspired by the clipped precision in Eq (4), we propose a clipped recall to measure how many words from the selected sentence xjcx^{c}_{j} are still covered in the refined sentence,

(6) an=gnxjcIDF(gn)min[Countclip(gn),Countx(gn)]gnxjcIDF(gn)Countclip(gn)a_{n}=\frac{\sum_{g^{n}\in x^{c}_{j}}IDF(g^{n})\cdot min[Count_{clip}(g^{n}),Count_{x}(g^{n})]}{\sum_{g^{n}\in x^{c}_{j}}IDF(g^{n})\cdot Count_{clip}(g^{n})}

where Countclip(gn)Count_{clip}(g^{n}) is the clipped count of n-gram gng^{n} towards the references like in BLEU, and Countx(gn)Count_{x}(g^{n}) is the count of gng^{n} in the refined explanation xx. In other words, the denominator is the prototype’s overlap with the target references and the numerator is the overlap among the prototype, references, and the final explanation. We did not use classical recall definition because it would reward the refiner to retain the entire prototype. We only encourage the refiner to keep the n-grams that are actually presented in the references. We compute the refiner’s intrinsic reward by aggregating the clipped recall over different n-grams πref(x,xjc)=exp(n=1Nwnlogan)\pi_{ref}(x,x^{c}_{j})=\exp\big{(}\sum^{N}_{n=1}w_{n}\log a_{n}\big{)}. We did not provide this reward to the extractor, because it biases the extractor to short and generic candidates which are easier for the refiner to cover.

With the hierarchical intrinsic rewards introduced for each component, we can optimize Eq (2) by policy gradient as

ΘJ[λ1π(x|u,c)+λ2πref(x,xjc)]ΘlogPref(x|xjc,rcu,Ωu)+[λ3π(x|u,c)+λ4πext(xjc)]ΘlogPext(xjc|rcu,Ωu)\begin{split}\nabla_{\Theta}J\approx&[\lambda_{1}\pi(x|u,c)+\lambda_{2}\pi_{ref}(x,x^{c}_{j})]\nabla_{\Theta}\log P_{ref}(x|x^{c}_{j},r^{u}_{c},\Omega_{u})\\ &+[\lambda_{3}\pi(x|u,c)+\lambda_{4}\pi_{ext}(x^{c}_{j})]\nabla_{\Theta}\log P_{ext}(x^{c}_{j}|r^{u}_{c},\Omega_{u})\end{split}

where λ1\lambda_{1} to λ4\lambda_{4} are coefficients to adjust the importance of each reward, and Θ\Theta stands for the model parameters in CompExp.

3.4. Model Training

The whole model training process can be organized into two steps: pre-training and fine-tuning. The pre-training step aims to bootstrap the extractor and refiner independently. To prepare the extractor to recognize the comparative relationships among sentences, we treat every observed review sentence as the extraction target and train the extractor to maximize its negative log-likelihood with regard to the corresponding user and item profiles.

It is important to pre-train the refiner as a generative language model, because it would be very inefficient to learn all the natural language model parameters only through the end-to-end training. However, we do not have any paired sentences to pre-train the refiner. We borrowed the method introduced in (Guu et al., 2018; Weston et al., 2018) to manually craft such pairs. Specifically, for every sentence, we compute its cosine similarity against all other sentences in the same item profile in the latent embedding space, and select the most similar one to pair with. Then we use this dataset to pre-train the refiner with negative log-likelihood loss.

In the fine-tuning stage, we concatenate the pre-trained layers and conduct the end-to-end training with policy gradient. To make the policy gradient training more resilient to variance and converge faster, it is important to have a baseline to update the model with reward advantages instead of using the rewards directly. We apply Monte Carlo sampling in both extractor and refiner to have multiple explanations, and use their mean rewards as the baseline.

4. Experimental Evaluations

We demonstrate empirically that CompExp can generate improved explanations compared to state-of-the-art explainable recommendation algorithms. We conduct experiments on two different recommendation scenarios: RateBeer reviews with single-ratings (McAuley et al., 2012) and TripAdvisor reviews with multi-aspect ratings (Wang et al., 2010).

4.1. Experiment Setup

As our solution only focuses on explanation generation, it can be applied to any recommendation algorithm of choice. In our experiments, we directly use the ground-truth review ratings as the recommendation score to factor out any deviation or noise introduced by specific recommendation algorithms. For completeness, we also empirically studied the impact from input ratings if switched to a real recommendation algorithm’s predictions.

Table 1. Summary of the processed datasets.
Dataset # Users # Items # Reviews  Rating Range
RateBeer 6,566 19,876 2,236,278 0 - 20
TripAdvisor 4,954 4,493 287,879 1 - 5

4.1.1. Data Pre-Processing

In the RateBeer dataset, we segment each review into sentences, and label them with the overall ratings from their original reviews. In the TripAdvisor dataset, there are separate ratings for five aspects including service, room, location, value and cleanliness. Therefore, each TripAdvisor review is expected to be a mix of a user’s opinions on these different aspects about the item. We segment sentences in a TripAdvisor review to different aspects using the boot-strapping method from (Wang et al., 2010) and assign resulting sentences the corresponding aspect ratings. These two datasets evaluate CompExp under different scenarios: overall opinion vs., aspect-specific opinion. We also adopt the recursive filtering (Wang et al., 2018b) to alleviate the data sparsity. The statistics of the processed datasets are summarized in Table 1.

4.1.2. Baselines

We compared with three explainable recommendation baselines that generate natural language explanations, covering both extraction-based and generation-based solutions.

  • -

    NARRE: Neural Attentional Regression model with Review-level Explanations (Chen et al., 2018). It is an extraction-based solution. It learns the usefulness of the existing reviews through attention and selects the most attentive reviews as the explanation.

  • -

    NRT: Neural Rating and Tips Generation (Li et al., 2017). It is a generation-based solution. It models rating regression and content generation as a multi-task learning problem with shared latent space. Content is generated from its neural language model component.

  • -

    SAER: Sentiment Aligned Explainable Recommendation (Yang et al., 2021). This is another generation-based solution using multi-task learning to model rating regression and explanation generation. But it focuses specifically on the sentiment alignment between the predicted rating and generated explanation.

We include three variants of CompExp to better demonstrate the effect of each component in it:

  • -

    CompExp-Ext: the extractor of our solution. It directly uses the selected sentences as explanations without any refinement. This variant helps us study how the extractor works and also serves as a fair counterpart for the other extraction-based baseline.

  • -

    CompExp-Pretrain: our model with pre-training only, which is a simple concatenation of the separately trained extractor and refiner without joint training. We compare it with CompExp to show the importance of end-to-end policy gradient training.

  • -

    CompExp-BLEU: our model trained with BLEU instead of IDF-BLEU. We create this variant to demonstrate the flaws of using BLEU to evaluate the quality of generated explanations.

Table 2. Explanation quality evaluated under IDF-BLEU, BLEU, average sentence length, average IDF per word, rep/l, seq_rep_2, feature precision and recall on RateBeer and TripAdvisor datasets. Bold numbers are the best of the corresponding metrics with p-value ¡ 0.05.
Model IDF-BLEU BLEU Avg Length IDF/word rep/l seq_rep_2 Feature
1 2 4 1 2 4 precision recall
RateBeer
Human / / / / / / 11.13 2.45 0.0535 0.0015 / /
NARRE 17.00 5.18 1.29 30.22 9.90 3.58 11.50 2.43 0.0643 0.0013 0.2217 0.0722
NRT 30.38 16.30 5.80 48.22 25.28 10.03 10.43 2.09 0.1123 0.0240 0.4563 0.1320
SAER 31.79 16.02 5.71 49.08 26.87 10.59 10.71 1.93 0.1146 0.0223 0.4751 0.1347
CompExp-Ext 24.86 11.72 2.99 38.54 18.74 5.98 12.10 2.36 0.0420 0.0010 0.3092 0.0929
CompExp-Pretrain 27.59 13.44 4.19 44.93 21.53 7.95 10.55 2.07 0.1448 0.0381 0.3922 0.1123
CompExp-BLEU 23.20 14.55 4.70 53.45 32.42 11.62 7.09 1.83 0.0266 0.0006 0.4025 0.1173
CompExp 32.36 19.55 6.95 49.14 29.63 11.41 10.52 2.16 0.0572 0.0057 0.4796 0.1383
TripAdvisor
Human / / / / / / 12.85 2.45 0.0604 0.0021 / /
NARRE 11.97 3.43 1.59 20.45 6.23 3.38 13.17 2.41 0.0641 0.0022 0.1733 0.1258
NRT 16.19 7.50 2.48 30.62 13.07 5.11 10.22 1.81 0.1277 0.0135 0.2939 0.1866
SAER 16.37 7.65 2.35 31.20 13.51 4.94 10.08 1.71 0.1361 0.0141 0.3178 0.1961
CompExp-Ext 13.52 4.25 1.14 22.12 7.30 2.66 14.70 2.39 0.0726 0.0037 0.2218 0.1553
CompExp-Pretrain 14.50 6.11 1.99 27.14 11.12 4.32 10.79 1.92 0.1177 0.0250 0.2736 0.1597
CompExp-BLEU 17.04 7.39 2.04 32.67 14.66 5.53 10.77 1.76 0.1597 0.0277 0.2332 0.1637
CompExp 21.35 8.01 2.16 31.70 12.23 4.16 13.35 2.12 0.0654 0.0053 0.3155 0.1930

4.2. Quality of Generated Explanations

To comprehensively study the quality of generated explanations, we employ different types of performance metrics, including IDF-BLEU-{1, 2, 4}, BLEU-{1, 2, 4}, average sentence length, average IDF per word, rep/l and seq_rep_2, and feature precision & recall. Both rep/l and seq_rep_2 are proposed in (Welleck et al., 2019) to evaluate content repetition and higher values mean the content is more repetitive. Features are items’ representative attributes that users usually care the most (Wang et al., 2018b; Xian et al., 2021; Yang et al., 2021), e.g., “pool” in Figure 1. The precision and recall measure if features mentioned in the generated explanations also appear in the user’s ground-truth review. We also include ground-truth review sentences as a reference baseline (labeled as “Human”) to study the differences between human and algorithm generated content. The results are reported in Table 2.

4.2.1. IDF-BLEU over BLEU.

While CompExp-BLEU topped every BLEU category on both datasets, CompExp also led almost all IDF-BLEU categories. This shows the effectiveness of our model design and the importance of directly optimizing the target evaluation metrics. To understand whether IDF-BLEU is a better metric than BLEU in evaluating the generated explanations, we should consider how the “ground-truth” content from real users look like, e.g., their average length and IDF/word, which suggest how much information is usually contained in a user-written sentence. As we can clearly notice that Avg Length and IDF/word in CompExp-BLEU are much smaller than Human. This suggests simply optimizing BLEU led to much shorter and less informative content. This follows our discussion before: BLEU encourage a model to generate less words and abuse common words to achieve high n-gram precision. CompExp-BLEU’s low feature precision and recall also reflect its weakness in providing informative content. Therefore, the witnessed “advantages” of CompExp-BLEU in BLEU most likely come from shorter and more generic sentences, instead of really being closer to the ground-truth content.

4.2.2. Advantages of CompExp.

There is clear performance gap between the extraction-based solutions (NARRE, CompExp-Ext) and generation-based ones (NRT, SAER, CompExp). While generation-based solutions largely outperformed extraction-based ones in content overlapping with ground-truth (IDF-BLEU, BLEU, feature precision and recall), they were generally very different from human writings in terms of sentence length, use of rare words (IDF/word), and content repetition (rep/l, seq_res_2). The extraction-based solutions use content provided by human, but they are limited to the existing content. The generation-based solutions customize content for each recommendation, but suffer from common flaws of generative models, e.g., short, dull, and repetitive. Among all the models, CompExp achieved the best balance among all metrics. It significantly exceeded all baselines in terms of IDF-BLEU-{1,2} and its BLEU was only behind CompExp-BLEU. Its feature precision and recall are competitive with SAER while leading the rest, though SAER enjoys additional advantage from predefined feature pool of each item as input. As a generation-based model, CompExp largely improved the average length, word rarity, and reduced repetition over NRT and SAER. The only exception is that CompExp-BLEU was less repetitive in RateBeer, but it is mainly because its explanations were very short in general.

4.2.3. Ablation Study.

Though CompExp performed well as a whole, it is inspiring to study if each component works as expected in the extract-and-refine workflow. First, both CompExp-Ext and NARRE are extraction-based with the same candidate pool, but CompExp-Ext showed obvious advantage under most categories of IDF-BLEU and BLEU. It suggests our extractor alone can act as a competent solution where generation-based models do not fit, e.g., real-time applications requiring minimum response time. The comparison between CompExp-Ext and CompExp-Pretrain demonstrates that the refiner is able to leverage the gradient direction to improve the prototypes, even when the prototypes are given by an extractor that has not been trained jointly with the refiner. At last, there are huge gaps in all metrics between CompExp-Pretrain and CompExp in both datasets. It is obvious that our reward design is beneficial to both quality and diversity of the generated explanations.

4.2.4. Comparativeness.

To verify if the generated explanations by CompExp capture the comparative ranking of items, we study its its output’s sensitivity to the input recommendation ratings. As a starting point, the ground-truth explanation perfectly aligns with the recommendation ranking, which is derived from the ground-truth rating. If the generated explanation carries the same ranking of item, the generated content should be close to the ground-truth content. As a result, if we manipulate the input recommendation scores of items, the generated explanations should start to deviate. The further we push the rankings apart, the further the generated explanation should be pushed away from the ground-truth explanation. We use IDF-BLEU and BLEU to measure the content similarity and perturb the recommendation ratings with Gaussian noise. As shown in Figure 3(a), all IDF-BLEU and BLEU metrics keep decreasing with the increasing amount of perturbation. In other words, even if it is for the same user and same set of items, with different recommendation scores assigned, CompExp would generate different explanations to explain their relative ranking.

4.2.5. Predicted Ratings.

Motivated by the findings in Figure 3(a), we further study how CompExp is influenced by a real recommendation algorithm’s predicted ratings. We employed the neural collaborative filtering (He et al., 2017) and used its predicted ratings in CompExp’s training and testing. The result is plotted in Figure 3(b). Compared with previous randomly perturbed ratings, the predicted ratings bring very limited changes to the explanations. This confirms our experiment results based on ground-truth ratings can fairly represent CompExp’s performance in real-world usage scenarios.

Refer to caption
(a)
Refer to caption
(b)
Figure 3. (a) Impact of noise in recommendation ratings on BLEU and IDF-BLEU. (b) Change in BLEU and IDF-BLEU with algorithm’s predicted ratings.

5. User Study

We have three research questions to answer in user study: 1) does users’ judgement toward explanation quality aligns more with IDF-BLEU than BLEU; 2) do users find our comparative explanations more helpful than the baselines’; and 3) can users better perceive the comparative ranking from our explanations than the baselines’. To answer these three research questions, we design two user study tasks based on RateBeer dataset using Amazon Mechanical Turk.

The first task studies the first two research questions together. Specifically, we shuffle explanations from different models about the same recommended item and ask the participants to compare them, and then select the most helpful ones. To help participants evaluate the explanation quality, we include the original user review as the item description, towards which they can judge if the explanation are accurate or informative. For each recommended item, we ask participants to answer the following question after reading its description and candidate explanations:

  • “Which of the following explanations best describe the characteristics of the given beer and help you the most to understand why you should pay attention to the recommendation?”

In this experiment, we collected 660 user responses.

The results are presented in Table 3 and 4. In Table 3, we used Cohen’s kappa coefficient to compare IDF-BLEU and BLEU’s agreement with users’ responses. For each test case, we pair explanations that the participants chose as helpful with the rest to form a set of explanation pairs. Then we use IDF-BLEU-{1,2,4} and BLEU-{1,2,4} to identify the helpful one in each pair. The kappa coefficient shows that IDF-BLEU aligns significantly better with users’ judgment in all three subcategories under paired t-test. Table 4 shows the helpfulness vote on each model and the paired t-test results of CompExp against other baselines. The helpfulness vote on CompExp is significantly higher than others, which suggests strong user preference over its generated explanations.

Table 3. Cohen’s kappa coefficient of explanation quality between the human judgements and BLEU & IDF-BLEU.
1 2 4
κ\kappa BLEU 0.2936 0.3114 0.2814
IDF-BLEU 0.3452 0.3396 0.3152
Paired t-test 0.0001 0.0094 0.0071
Table 4. Up-vote rate of explanations’ helpfulness.
CompExp SAER NRT NARRE
Up-vote Rate 43.79% 37.27% 35.61% 30.61%
Paired t-test / 0.0182 0.009 0

The second task addresses the last research question, i.e., if a user is able to perceive the ranking of recommended items from the explanations. In this task, we randomly paired items of different ratings and asked participants to identify which item is better by reading the provided explanations. We then evaluated the agreement rate between participants’ choices and the actual ranking. In particular, given the explanations of a model, the participants were required to answer the following question:

  • “After reading the explanations for recommended items, which item would you like to choose? You are expected to judge the quality of the items based on the provided explanations.”

We chose SAER and NRT as baselines. Besides, we also include the ground-truth sentences from the actual user reviews as a reference. We collected 200 responses for each model.

Table 5. Agreement rate between actual ranking and the users perceived ranking of paired items based on the provided explanations.
GT CompExp SAER NRT
Agreement Rate 72.29% 57.27% 56.25% 53.14%

Table 5 reports the agreement rates between the actual ranking and the ranking perceived by the participants. CompExp’s agreement rate is slightly higher than NRT and SAER, but it is far below the Ground-Truth. The Ground-Truth’s high agreement rate quantitatively confirms that the original user provided review sentences are highly comparative. This observation supports our choice of training the comparative explanation generation from paired user review sentences. And it also suggests there is still a performance gap in comparativeness for learning-based solutions to bridge. And an improved objective for optimization, e.g., include quantified pairwise comparativeness, might be a promising direction.

6. Conclusion and Future Work

In this paper, we studied the problem of comparative explanation generation in explainable recommendation. The objective of our generated explanations is to help users understand the comparative item rankings provided in a recommender system. We develop a neural extract-and-refine architecture to generate such comparative explanations, with customized metrics to penalize generic and useless content in the generated explanations. Both offline evaluations and user studies demonstrated the effectiveness of our solution.

This work starts a bright new direction in explainable recommendation. Our current solution only focuses on explanation generation, by assuming a perfect recommendation algorithm (i.e., we directly used the ground-truth opinion ratings in our experiments). It is important to improve our model by co-design with a real recommendation algorithm, whose recommendation score is expected to be noise and erroneous. In addition, we still heavily depend on existing review content to guide explanation generation. It will be more meaningful to introduce actual user feedback in this process, i.e., interactive optimization of explanation generation.

Acknowledgements.
We thank the anonymous reviewers for their insightful comments and suggestions. This work is partially supported by the National Science Foundation under grant SCH-1838615, IIS-1553568, and IIS-2007492, and by Alibaba Group through Alibaba Innovative Research Program.

References

  • (1)
  • Aggarwal et al. (2016) Charu C Aggarwal et al. 2016. Recommender systems. Vol. 1. Springer.
  • Ai et al. (2018) Qingyao Ai, Vahid Azizi, Xu Chen, and Yongfeng Zhang. 2018. Learning heterogeneous knowledge base embeddings for explainable recommendation. Algorithms 11, 9 (2018), 137.
  • Balabanović and Shoham (1997) Marko Balabanović and Yoav Shoham. 1997. Fab: content-based, collaborative recommendation. Commun. ACM 40, 3 (1997), 66–72.
  • Bilgic and Mooney (2005) Mustafa Bilgic and Raymond J Mooney. 2005. Explaining recommendations: Satisfaction vs. promotion. In Beyond Personalization Workshop, IUI, Vol. 5.
  • Cai et al. (2017) Renqin Cai, Chi Wang, and Hongning Wang. 2017. Accounting for the Correspondence in Commented Data. In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval. 365–374.
  • Cai et al. (2021) Renqin Cai, Jibang Wu, Aidan San, Chong Wang, and Hongning Wang. 2021. Category-aware collaborative sequential recommendation. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval. 388–397.
  • Chen et al. (2018) Chong Chen, Min Zhang, Yiqun Liu, and Shaoping Ma. 2018. Neural attentional rating regression with review-level explanations. In Proceedings of the 2018 World Wide Web Conference. 1583–1592.
  • Chen et al. (2021) Hongxu Chen, Yicong Li, Xiangguo Sun, Guandong Xu, and Hongzhi Yin. 2021. Temporal meta-path guided explainable recommendation. In Proceedings of the 14th ACM International Conference on Web Search and Data Mining. 1056–1064.
  • Chen et al. (2020) Tong Chen, Hongzhi Yin, Guanhua Ye, Zi Huang, Yang Wang, and Meng Wang. 2020. Try this instead: Personalized and interpretable substitute recommendation. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval. 891–900.
  • Chung et al. (2014) Junyoung Chung, Caglar Gulcehre, KyungHyun Cho, and Yoshua Bengio. 2014. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555 (2014).
  • Conneau et al. (2017) Alexis Conneau, Douwe Kiela, Holger Schwenk, Loïc Barrault, and Antoine Bordes. 2017. Supervised Learning of Universal Sentence Representations from Natural Language Inference Data. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing. 670–680.
  • Guu et al. (2018) Kelvin Guu, Tatsunori B Hashimoto, Yonatan Oren, and Percy Liang. 2018. Generating sentences by editing prototypes. Transactions of the Association for Computational Linguistics 6 (2018), 437–450.
  • He et al. (2015) Xiangnan He, Tao Chen, Min-Yen Kan, and Xiao Chen. 2015. Trirank: Review-aware explainable recommendation by modeling aspects. In Proceedings of the 24th ACM International on Conference on Information and Knowledge Management. 1661–1670.
  • He et al. (2017) Xiangnan He, Lizi Liao, Hanwang Zhang, Liqiang Nie, Xia Hu, and Tat-Seng Chua. 2017. Neural collaborative filtering. In Proceedings of the 26th international conference on world wide web. 173–182.
  • Herlocker et al. (2000) Jonathan L Herlocker, Joseph A Konstan, and John Riedl. 2000. Explaining collaborative filtering recommendations. In Proceedings of the 2000 ACM conference on Computer supported cooperative work. ACM, 241–250.
  • Holtzman et al. (2019) Ari Holtzman, Jan Buys, Li Du, Maxwell Forbes, and Yejin Choi. 2019. The curious case of neural text degeneration. arXiv preprint arXiv:1904.09751 (2019).
  • Ji and Shen (2016) Ke Ji and Hong Shen. 2016. Jointly modeling content, social network and ratings for explainable and cold-start recommendation. Neurocomputing 218 (2016), 1–12.
  • Karatzoglou et al. (2013) Alexandros Karatzoglou, Linas Baltrunas, and Yue Shi. 2013. Learning to rank for recommender systems. In Proceedings of the 7th ACM Conference on Recommender Systems. 493–494.
  • Koren et al. (2009) Yehuda Koren, Robert Bell, and Chris Volinsky. 2009. Matrix factorization techniques for recommender systems. Computer 42, 8 (2009), 30–37.
  • Li et al. (2019) Chenliang Li, Cong Quan, Li Peng, Yunwei Qi, Yuming Deng, and Libing Wu. 2019. A Capsule Network for Recommendation and Explaining What You Like and Dislike. In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval. 275–284.
  • Li et al. (2017) Piji Li, Zihao Wang, Zhaochun Ren, Lidong Bing, and Wai Lam. 2017. Neural rating regression with abstractive tips generation for recommendation. In Proceedings of the 40th International ACM SIGIR conference on Research and Development in Information Retrieval. 345–354.
  • Lundberg and Lee (2017) Scott M Lundberg and Su-In Lee. 2017. A unified approach to interpreting model predictions. In Advances in neural information processing systems. 4765–4774.
  • Luong et al. (2015) Minh-Thang Luong, Hieu Pham, and Christopher D Manning. 2015. Effective approaches to attention-based neural machine translation. arXiv preprint arXiv:1508.04025 (2015).
  • McAuley et al. (2012) Julian McAuley, Jure Leskovec, and Dan Jurafsky. 2012. Learning Attitudes and Attributes from Multi-Aspect Reviews. In Proceedings of the 2012 IEEE 12th International Conference on Data Mining (ICDM ’12). IEEE Computer Society, USA, 1020–1025.
  • McAuley et al. (2015) Julian McAuley, Rahul Pandey, and Jure Leskovec. 2015. Inferring networks of substitutable and complementary products. In Proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining. 785–794.
  • Papineni et al. (2002) Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. BLEU: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting on association for computational linguistics. Association for Computational Linguistics, 311–318.
  • Pennington et al. (2014) Jeffrey Pennington, Richard Socher, and Christopher D Manning. 2014. Glove: Global vectors for word representation. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP). 1532–1543.
  • Rendle (2010) Steffen Rendle. 2010. Factorization machines. In 2010 IEEE International Conference on Data Mining. IEEE, 995–1000.
  • Rendle et al. (2012) Steffen Rendle, Christoph Freudenthaler, Zeno Gantner, and Lars Schmidt-Thieme. 2012. BPR: Bayesian personalized ranking from implicit feedback. arXiv preprint arXiv:1205.2618 (2012).
  • Ribeiro et al. (2016) Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. 2016. ”Why should i trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining. 1135–1144.
  • Sarwar et al. (2001) Badrul Sarwar, George Karypis, Joseph Konstan, and John Riedl. 2001. Item-based collaborative filtering recommendation algorithms. In Proceedings of the 10th international conference on World Wide Web. 285–295.
  • Sinha and Swearingen (2002) Rashmi Sinha and Kirsten Swearingen. 2002. The role of transparency in recommender systems. In CHI’02 extended abstracts on Human factors in computing systems. ACM, 830–831.
  • Sun et al. (2020) Peijie Sun, Le Wu, Kun Zhang, Yanjie Fu, Richang Hong, and Meng Wang. 2020. Dual Learning for Explainable Recommendation: Towards Unifying User Preference Prediction and Review Generation. In Proceedings of The Web Conference 2020. 837–847.
  • Sutskever et al. (2014) Ilya Sutskever, Oriol Vinyals, and Quoc V Le. 2014. Sequence to sequence learning with neural networks. In Advances in neural information processing systems. 3104–3112.
  • Tao et al. (2019) Yiyi Tao, Yiling Jia, Nan Wang, and Hongning Wang. 2019. The FacT: Taming Latent Factor Models for Explainability with Factorization Trees. In Proceedings of the 42Nd International ACM SIGIR Conference on Research and Development in Information Retrieval. ACM, New York, NY, USA, 295–304.
  • Truong and Lauw (2019) Quoc-Tuan Truong and Hady Lauw. 2019. Multimodal Review Generation for Recommender Systems. In The World Wide Web Conference. 1864–1874.
  • Vezhnevets et al. (2017) Alexander Sasha Vezhnevets, Simon Osindero, Tom Schaul, Nicolas Heess, Max Jaderberg, David Silver, and Koray Kavukcuoglu. 2017. Feudal networks for hierarchical reinforcement learning. In International Conference on Machine Learning. PMLR, 3540–3549.
  • Wang et al. (2010) Hongning Wang, Yue Lu, and Chengxiang Zhai. 2010. Latent aspect rating analysis on review text data: a rating regression approach. In Proceedings of the 16th ACM SIGKDD international conference on Knowledge discovery and data mining. 783–792.
  • Wang et al. (2018b) Nan Wang, Hongning Wang, Yiling Jia, and Yue Yin. 2018b. Explainable recommendation via multi-task learning in opinionated text data. In The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval. 165–174.
  • Wang et al. (2018a) Xiting Wang, Yiru Chen, Jie Yang, Le Wu, Zhengtao Wu, and Xing Xie. 2018a. A reinforcement learning framework for explainable recommendation. In 2018 IEEE International Conference on Data Mining (ICDM). IEEE, 587–596.
  • Welleck et al. (2019) Sean Welleck, Ilia Kulikov, Stephen Roller, Emily Dinan, Kyunghyun Cho, and Jason Weston. 2019. Neural text generation with unlikelihood training. arXiv preprint arXiv:1908.04319 (2019).
  • Weston et al. (2018) Jason Weston, Emily Dinan, and Alexander H Miller. 2018. Retrieve and refine: Improved sequence generation models for dialogue. arXiv preprint arXiv:1808.04776 (2018).
  • Wu et al. (2020) Jibang Wu, Renqin Cai, and Hongning Wang. 2020. Déjà vu: A contextualized temporal attention mechanism for sequential recommendation. In Proceedings of The Web Conference 2020. 2199–2209.
  • Xian et al. (2019) Yikun Xian, Zuohui Fu, Shan Muthukrishnan, Gerard De Melo, and Yongfeng Zhang. 2019. Reinforcement knowledge graph reasoning for explainable recommendation. In Proceedings of the 42nd international ACM SIGIR conference on research and development in information retrieval. 285–294.
  • Xian et al. (2021) Yikun Xian, Tong Zhao, Jin Li, Jim Chan, Andrey Kan, Jun Ma, Xin Luna Dong, Christos Faloutsos, George Karypis, Shan Muthukrishnan, et al. 2021. EX3: Explainable Attribute-aware Item-set Recommendations. In Fifteenth ACM Conference on Recommender Systems. 484–494.
  • Yang et al. (2021) Aobo Yang, Nan Wang, Hongbo Deng, and Hongning Wang. 2021. Explanation as a Defense of Recommendation. In Proceedings of the 14th ACM International Conference on Web Search and Data Mining. 1029–1037.
  • Zhang and Chen (2020) Yongfeng Zhang and Xu Chen. 2020. Explainable recommendation: A survey and new perspectives. Foundations and Trends® in Information Retrieval 14, 1 (2020), 1–101.
  • Zhang et al. (2014) Yongfeng Zhang, Guokun Lai, Min Zhang, Yi Zhang, Yiqun Liu, and Shaoping Ma. 2014. Explicit factor models for explainable recommendation based on phrase-level sentiment analysis. In Proceedings of the 37th international ACM SIGIR conference on Research & development in information retrieval. 83–92.
Table 6. Case study of the explanations generated by different models.
Model Sample 1 Sample 2
Human aroma of caramel, cherry, raisins, and florals. pours clear yellow body with a small white head.
NARRE the finish is dry and ashy. not bad, if one is looking for a refreshing, light wheat beer.
NRT flavor of chocolate, roasted malt, and light smoke. the beer is a hazy yellow-orange color.
SAER aroma of caramel, caramel, and citrus. medium body, watery texture, and carbonation.
CompExp-Ext sweet aroma with toasted malt, caramel and alcohol notes. pours a hazy golden with a small white head.
CompExp aroma of caramel, malt, and alcohol. pours a hazy yellow body with a small white head.

Appendix A Model Implementation Details

In the section, we will share our technical choices of some important components and hyper-parameter values in CompExp.

CompExp’s extractor adopts a single text encoder to obtain universal sentence representations for reviews from both user and item profile. The text encoder’s architecture follows the self-attentive network presented in (Conneau et al., 2017), where the attention mechanism aggregates the hidden states of a bi-directional GRU. The GRU is of a single layer with hidden state size of 300. For the input word embeddings, we bootstrap their initial values with GloVe 6B of 300 dimensions (Pennington et al., 2014) and allow them to be further updated during training.

As we discussed before, our proposed solution is able to handle both continuous and discrete ratings, but in this work, we assume recommendation ratings are discrete. Therefore, our implementation applies a rating embeddings to map the one-hot vector of rating difference into its latent representation. These rating embeddings are randomly initialized and learned through the back-propagation during the training process. We set the embedding size to 16. For the following latent space transformation f(xiu,Δriu)hif(x^{u}_{i},\Delta r^{u}_{i})\to h_{i}, we use a 2-layer MLP with Tanh as the activation function, whose intermediate and final output sizes are both 300.

The final text decoder inside CompExp’s refiner is another single layer GRU with hidden size of 300. The text encoder from the extractor and this decoder together actually forms a sequence-to-sequence model (Sutskever et al., 2014) whose input is the chosen prototype sentence. So our extract-and-refine process can be viewed as multiple sequence encoders run in parallel, while only one of them can connect to the sequence decoder. Additionally, the text decoder also adopts the attention layer proposed in (Luong et al., 2015). We find that paying attention to the extracted prototype during the refining process is beneficial to the clipped recall defined in Eq (6).

Since this work focuses on studying the problem of comparative explanation, the above techniques are enough for us to demonstrate the effectiveness of our proposed solution. The architecture of CompExp itself does not hold any assumptions about the implementation of the discussed sub-components. They can be replaced with other state-of-the-art models to further boost the performance.

Appendix B Model Training Details

In the section, we will discuss the techniques we used to train CompExp and the corresponding hyper-parameters.

To train the extractor, we need to batch multiple user profiles and item profile respectively. However, the sizes of the profiles vary a lot. When we batch the profiles with all their reviews, the batch will end with many paddings to ensure every profile within the batch has the same size. These paddings waste lots of computing resources and heavily slow down the training process. So instead of using all the reviews, we define a max limit. For a profile larger than the limit, we will randomly sample a subset based on the limit. Larger limit usually leads to better training results since the model have more references and candidates to leverage. We set the limit to 10 for both user and item based on our computing capacity and we also found the improvement beyond it is marginal.

The value of κ\kappa from Eq (3) is critical to the training since it balances the exploration and exploitation in the policy gradient. Smaller values flatten the extraction distribution and hence force the model to explore more extraction candidates, but this tends to delay the convergence and cause very unstable results. On the other hand, larger values concentrate the distribution and reduce the search space, but then the model may miss more appropriate candidates and lose the meaning of the joint training. Based on our tests, we find 3 is the most balanced value.

IDF-BLEU is the main reward in the policy gradient training, but we used unconventional n-gram weights. Following the original design of BLEU, IDF-BLEU keeps the individual weight for each type of n-gram precision, i.e., wnw_{n} in Eq (5). The weights we used for unigram to 4-gram are {0.8, 0.2, 0, 0}. Usually, the weights are equally distributed among all the available n-grams. For example, BLEU-2 has the weight of 0.5 for both unigram and bigram; similarly, IDF-BLEU-4 applies a unified weight of 0.25 from unigram to 4-gram. However, the precision of different n-grams are not always compatible with each other as objectives and the model has to make trade-offs. For example, we found using IDF-BLEU-4 as reward sacrifice unigram and bigram precision in exchange for 4-gram precision. As a result, it only slightly benefits the IDF-BLEU-4 in evaluation but leads IDF-BLEU-{1,2} to decline. Therefore, we decided this customized weights to only focus on unigram and bigram. There are two reasons. First, correctly generating trigrams and 4-grams in explanations is quite difficult so such overlaps will not be very frequent anyway. Instead of betting for some dull 4-grams, it is more valuable to cover the interested features which are often just unigrams. Second, higher precision of unigram and bigram still contribute to IDF-BLEU-4 in evaluation according to its definition. But they may not always compensate the negligence of other n-grams. This explains why CompExp has a less competitive IDF-BLEU-4 in TripAdvisor meanwhile leading the rest in Table 2.

Appendix C Case Study

Groups of example explanations generated by CompExp and other baselines are shown in Table 6. The ground-truth explanations are given for reference denoted as Human. Comparing NARRE and CompExp-Ext shows the value of modeling comparativeness in users provided historical content. Sentences extracted by CompExp-Ext are much closer to the ground-truth than NARRE’s. Comparing CompExp-Ext and CompExp shows the effectiveness our rewriting module in improving the explanation quality, especially in writing style and wording. For example, in Sample 1, the extracted explanation correctly covers the attribute “aroma” and “caramel”, but its sentence structure is different from the ground-truth’s. The refined explanation keeps the two correct attributes and improves the sentence structure. In Sample 2, while the extractor picks a sentence almost the same as the ground-truth, the refiner further changes the word “golden” to “yellow”, which better reflects the user’s preference in wording. However, both samples also suggest our refiner can be further improved in personalized feature revision. For example, in Sample 2, the end explanation inherits ”hazy” from the extracted prototype while the item looks ”clear” instead for the target user. Same for ”malt” and ”alcohol” vs., ”cherry” and ”raisins” in Sample 1. Obviously, these features are subjective and the target user hold a different opinion from the author of the extracted sentence. It would be a promising future direction to better personalize subjective features while still maintain the relevance and faithfulness to other objective facts given by the extraction.