This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Dual Task Framework for Improving Persona-grounded Dialogue Dataset

Minju Kim1, Beong-woo Kwak1, Youngwook Kim1, Hong-in Lee1
Seung-won Hwang2 and Jinyoung Yeo1
The authors contribute equally to this paper.Corresponding author (Email: jinyeo@yonsei.ac.kr)
Abstract

This paper introduces a simple yet effective data-centric approach for the task of improving persona-conditioned dialogue agents. Prior model-centric approaches unquestioningly depend on the raw crowdsourced benchmark datasets such as Persona-Chat. In contrast, we aim to fix annotation artifacts in benchmarking, which is orthogonally applicable to any dialogue model. Specifically, we augment relevant personas to improve dialogue dataset/agent, by leveraging the primal-dual structure of the two tasks, predicting dialogue responses and personas based on each other. Experiments on Persona-Chat show that our approach outperforms pre-trained LMs by an 11.7 point gain in terms of accuracy.

Introduction

In personalized dialogue agents, persona grounding has been a long-standing goal to improve both human-likeness and dialogue consistency. For example, when given a profile description such as “I am a doctor.” and “I don’t eat meat”, dialogue agents aim to plausibly respond to dialogue context while endowing the machine with the persona, e.g., “I am now working at hospital” and “I went vegan.” respectively. To learn and evaluate such grounding, recent research (Kim, Kim, and Kim 2020; Li et al. 2020; Song et al. 2020; Zhang et al. 2019) has proposed to encode personas based on pre-trained language models (PLMs) (Devlin et al. 2019; Liu et al. 2019; Radford et al. 2019), and many persona-conditioned dialogue benchmarks have been released such as Persona-Chat (Zhang et al. 2018), where crowdworkers role-play following the given description of personas to populate dialogues.

In spite of such recent significant progress, we argue that there is much room for improving persona-grounded dialogue agents in the data-centric view. Specifically, as crowdworkers often miss out on stating detailed experience and knowledge, it is reported that several linguistic biases exist in the persona-grounded dialogue datasets, for example, crowdworkers tend to reuse some terms of persona descriptions to script dialogue utterances, e.g., “I am a doctor.”, in which words trivially overlap to a persona sentence, or merely script utterances as a paraphrase, e.g., “I work as a doctor.” or “I don’t eat any meat.”.

Refer to caption
Figure 1: Illustration of the primal-dual task pipeline

Such annotation artifacts can be a bottleneck for learning to ground more engaging utterances and adversarially learning not to make contradictory utterances to personas. In practice, dialogue agents hardly respond in an engaging manner, e.g., “I am now working at hospital.”, loosening the linguistic ties from the given persona (Ghazarian et al. 2021). Also, on a challenging question such as “Are you vegan?” and “Do you like steak?”, dialogue agents may violate consistency (Welleck et al. 2018), e.g., responding “No” and “Yes” respectively. A straightforward way to mitigate these phenomena is to manually fix the dataset with additional annotation costs (Bowman and Dahl 2021), by asking another crowdworkers to rewrite/augment the dialogue utterances, but it is too expensive on a large scale.

As motivated above, in this work, we propose to automatically improve the persona-conditioned dialogue dataset to learn better persona-grounded dialogue agents. As illustrated in Figure 1, our key idea is to infer and add new personas from a dialogue, e.g., “I am a doctor.\mapsto I work at hospital or “I don’t eat any meat.\mapsto I am vegan. For that, we present a novel dual task framework, where the primal task Persona-Chat is learning persona-conditioned dialogue models (i.e., predicting utterances based on persona), whose personas are augmented by the dual yet secondary task, namely Persona-Link (i.e., predicting personas for utterances reversely). Our framework first learns the Persona-Link models, which use dialogue contexts in Persona-Chat as a query to augment its relevant personas, then learns the Persona-Chat models with the augmented personas for better persona grounding.

Straightforwardly, reversing the Persona-Chat dataset can provide the utterance-to-persona alignments as linking supervisions to train the Persona-Link models. However, using naive closed alignments from individual dialogue episodes may inherit the linguistic bias from Persona-Chat, which disqualifies the ultimate role of debiasing Persona-Chat. To tackle this challenge, we first adapt ideas from semantic matching (Wu et al. 2020) to our dual task, which enables augmenting open alignments out of individual dialogue episodes, i.e., less linguistically-biased alignments from the whole Persona-Chat corpus. Then, furthermore, we also leverage commonsense expansion (Majumder et al. 2020), which enables augmenting more open alignments out of the Persona-Chat corpus itself, in a systematic way.

Our main contributions are summarized as follows: (1) We propose an iterative framework of primal-dual tasks to debias the Persona-Chat dataset/model without any human effort. (2) We automate the learning of Persona-Link models from the Persona-Chat dataset, as a desired form of debiasing Persona-Chat. (3) Our extensive experiments validate that, along with linked personas, the response accuracy significantly increases by 11.7% point on Persona-Chat compared to that of using the raw dataset.

Primal-Dual Task Framework

Primal Task: Persona-Chat

The goal of the Persona-Chat task (Zhang et al. 2018) is to personalize dialogue agents by grounding their utterances to the given personas. The original dataset involves dialogues between pairs of speakers: each speaker is given a hypothetical profile, which is a few persona sentences that describe a character they will imitate, and is instructed to get to know the other. Formally, a profile is defined as a set of persona sentences 𝒫={p1,,pm}\mbox{${\cal P}$}=\{p_{1},...,p_{m}\} and a dialogue is defined as a set of multi-turn utterances 𝒰={u1,,un}\mbox{${\cal U}$}=\{u_{1},...,u_{n}\}, and a Persona-Chat dataset 𝒟Chat={(𝒫i,𝒰i)}i=1N\mathcal{D}_{\textsc{Chat}}=\{(\mbox{${\cal P}$}_{i},\mbox{${\cal U}$}_{i})\}_{i=1}^{N} where NN is the number of dialogues. Based on 𝒟Chat\mathcal{D}_{\textsc{Chat}}, when given persona sentences 𝒫{\cal P}, dialogue history 𝒰{\cal U}, and response space 𝕌\mathbb{U}, the primal task aims at finding a model f:(𝒫,𝒰)𝕌f:(\mbox{${\cal P}$},\mbox{${\cal U}$})\mapsto\mathbb{U}.

f(𝒫,𝒰;θChat)=Δargmaxu𝕌P(u|𝒫,𝒰;θChat)f(\mbox{${\cal P}$},\mbox{${\cal U}$};\theta_{\textsc{Chat}})\overset{\Delta}{=}\underset{u\in\mathbb{U}}{\operatornamewithlimits{argmax\ }}P(u|\mbox{${\cal P}$},\mbox{${\cal U}$};\theta_{\textsc{Chat}}) (1)

where θChat\theta_{\textsc{Chat}} is the parameters to be learned for model fθf_{\theta}. In our setting, we adopt the response selection task (Humeau et al. 2020). As response space 𝕌\mathbb{U}, the model has to pick the correct response from a set of 20 choices, where the remaining 19 were randomly chosen utterances from the evaluation set. Note that in a final system, however, one would retrieve from the entire training set of over 100k utterances, but this is avoided for speed reasons in common evaluation setups (Wolf et al. 2019; Humeau et al. 2020).

Dual Task: Persona-Link

Following the primal-dual structure, the Persona-Link task can be defined as: given an arbitrary dialogue utterance uu, the output of a linking model gg is its referent persona description in persona space \mathbb{P}. That is, the dual task aims at finding a linking model g:ug:u\mapsto\mathbb{P}.

g(u;θLink)=ΔargmaxpP(p|u;θLink)g(u;\theta_{\textsc{Link}})\overset{\Delta}{=}\underset{p\in\mathbb{P}}{\operatornamewithlimits{argmax\ }}P(p|u;\theta_{\textsc{Link}}) (2)

where θLink\theta_{\textsc{Link}} is the parameters to be learned for model gθg_{\theta}. We formalize the dual task as a variant of entity linking system, where an utterance is linked to an entry in the PKB \mathbb{P} of arbitrary size, being populated from the primal dataset 𝒟Chat\mathcal{D}_{\textsc{Chat}}, i.e., =i=1N𝒫i\mathbb{P}=\bigcup_{i=1}^{N}\mbox{${\cal P}$}_{i}. To learn model gθg_{\theta}, we adopt the same neural architecture of the primal task (i.e., Bi-encoder) (Humeau et al. 2020) as a base model with the cross-entropy loss.111For more flexible application, inspired by (Wu et al. 2020), we relax the collective linking in Eq. (2) into finer-level linking computing the utterance-to-persona score, i.e., p(pi|ui;θLink)p(p_{i}|u_{i};\theta_{\textsc{Link}}).

Input: Original data 𝒟Chat\mbox{${\cal D}$}_{\textsc{Chat}}
Output: Debiased data D~Chat\tilde{D}_{\textsc{Chat}}, Debiased model θ~Chat\tilde{\theta}_{\textsc{Chat}}
𝒟Link,𝒟~LinkReverseDataset(𝒟Chat)\mbox{${\cal D}$}_{\textsc{Link}},\tilde{\mbox{${\cal D}$}}_{\textsc{Link}}\leftarrow\text{ReverseDataset}(\mbox{${\cal D}$}_{\textsc{Chat}})      
θLinkTrain(gθ,𝒟Link)\theta_{\textsc{Link}}\leftarrow\text{Train}(g_{\theta},\mbox{${\cal D}$}_{\textsc{Link}})
θ~LinkTrain(gθ,𝒟~Link,θLink)\tilde{\theta}_{\textsc{Link}}\leftarrow\text{Train}(g_{\theta},\tilde{\mbox{${\cal D}$}}_{\textsc{Link}},\theta_{\textsc{Link}})
𝒟~ChatAugmentPersona(𝒟Chat,θ~Link)\tilde{\mbox{${\cal D}$}}_{\textsc{Chat}}\leftarrow\text{AugmentPersona}(\mbox{${\cal D}$}_{\textsc{Chat}},\tilde{\theta}_{\textsc{Link}}) θ~ChatTrain(fθ,D~Chat)\tilde{\theta}_{\textsc{Chat}}\leftarrow\text{Train}(f_{\theta},\tilde{D}_{\textsc{Chat}})
return  D~Chat,θ~Chat\tilde{D}_{\textsc{Chat}},\tilde{\theta}_{\textsc{Chat}}
Algorithm 1 Primal-Dual Task Framework
Persona-Link Debiased Persona-Chat
Refer to caption
Figure 2: Overall procedure of learning linking models at semantic and commonsense level for Persona-Link

Primal-Dual Task Pipeline

We introduce a new framework that exploits the duality (Xia et al. 2017) of Persona-Chat and Persona-Link tasks where the input (utterance) and output (persona) of the dual task are roughly the inverse of its primal task.

In the training phase of the primal task, our ultimate goal is to transform the original dialogue dataset 𝒟Chat\mbox{${\cal D}$}_{\textsc{Chat}} into the debiased dataset 𝒟~Chat\tilde{\mbox{${\cal D}$}}_{\textsc{Chat}} by augmenting new personas, then train the debiased dialogue model fθf_{\theta} from 𝒟~Chat\tilde{\mbox{${\cal D}$}}_{\textsc{Chat}}, which reduces the model dependence on linguistic bias between personas and utterances. Algorithm 1 describes its overall procedure. In the proposed framework, the linking model gθg_{\theta} should be learned in advance to augment plausible personas per a given utterance. However, the key challenge is collecting the supervisory information as the training data of gθg_{\theta} (denoted as 𝒟Link\mbox{${\cal D}$}_{\textsc{Link}} or 𝒟~Link\tilde{\mbox{${\cal D}$}}_{\textsc{Link}}), which is capable of injecting the debiasing ability to gθg_{\theta}. As illustrated in Figure 2, we mainly discuss this in the subsequent section.

Once an arbitrary linking model gθg_{\theta} is obtained as the desired debiasing function, we leverage gθg_{\theta} also in the inference phase of the primal task, not only in training time. Specifically, as illustrated in Figure 1, a new persona is interactively augmented per new response to ground the dialogue agents to debiased persona-response alignments. This procedure is repeated until the dialogue ends.

Reversing Persona-Chat to Persona-Link

A straightforward way to learn the dual task, Persona-Link, is to reverse input and output sides of the primal task each other (Wang et al. 2021; Su, Chuang, and Chen 2020; Zhu, Cao, and Yu 2020). However, in contrast to prior work, such a naive approach disqualifies the ultimate role of debiasing Persona-Chat since the Persona-Link data and model can inherit even the linguistic biases presented in the Persona-Chat dataset. Thus, the Persona-Link data should involve the desired asymmetric characteristics with Persona-Chat. We now present the three phases for learning Persona-Link models.

Phase 1: Out-dialogue Semantic Matching

Basically, as supervisory data for Persona-Link, we can collect the plausible alignments of persona and utterances in individual dialogue episodes of the Persona-Chat dataset, by using their semantic relationship. Specifically, we leverage natural language inference (NLI), a task of determining whether a hypothesis (e.g., persona) can be inferred from the given premise (e.g., utterance). The hypothesis sentence is classified into three categories: Entailment (true), Contradiction (false), and Neutral (undetermined). We adopt a RoBERTa model (Liu et al. 2019) trained on MNLI (Wang et al. 2019). By NLI, the minimal sampling to populate 𝒟Link\mbox{${\cal D}$}_{\textsc{Link}} is to consider the in-dialogue matching of persona-utterance pair (u,p)(u,p) as candidates, where uu and pp are derived from the same dialogue episode. However, as mentioned earlier, such (u,p)(u,p) pairs in individual dialogue episodes reportedly suffer from trivial word overlaps.

In contrast, we perform the out-dialogue matching of all possible combinations in 𝒟Chat\mbox{${\cal D}$}_{Chat}, which is stochastically less overlapped at lexical level. That is, such a simple strategy can contribute to mitigating the linguistic bias in the NLI-driven alignments, which we call seed Persona-Link dataset.

Definition 1

(Seed Persona-Link Dataset) Given 𝒟Chat={(𝒫j,𝒰j)}j=1N\mathcal{D}_{\textsc{Chat}}=\{(\mbox{${\cal P}$}_{j},\mbox{${\cal U}$}_{j})\}_{j=1}^{N}, a seed Persona-Link dataset can be defined as 𝒟Link={(ui,pi,yi)}i=1M\mbox{${\cal D}$}_{\textsc{Link}}=\{(u_{i},p_{i},y_{i})\}_{i=1}^{M} where uiu_{i} and pip_{i} are drawn from 𝒟Chat\mathcal{D}_{\textsc{Chat}}, i.e., uij=1N𝒰ju_{i}\in\bigcup_{j=1}^{N}\mbox{${\cal U}$}_{j} and pij=1N𝒫jp_{i}\in\bigcup_{j=1}^{N}\mbox{${\cal P}$}_{j}, and binary label yiy_{i} is captured by a NLI classifier. If (ui,pi)(u_{i},p_{i}) is inferred as Entailment, yi=1y_{i}=1, and yi=0y_{i}=0 otherwise.

Phase 2: Out-corpus Commonsense Expansion

Linguistic bias comes from the limited knowledge and experience of individual crowdworkers, which motivates the out-dialogue matching of personas and annotated utterances. Beyond the Persona-Chat corpus, if someone would have world-level knowledge, she may willingly annotate more engaging, less linguistically-biased utterances for a given persona. For example, utterance “I don’t eat any meat.” can involve commonsense attributes ‘I needed to be vegeterian’ or ‘I wanted to be healthy’, which enable drawing a new persona “I am vegan” from PKB, which are not captured by semantic matching (i.e., predicted as Neutral).

If a linking model is trained on sentence pairs of only similar semantics, such a potentially relevant mapping cannot be captured for generalization in inference time. Thus, since personas and utterances are instances of world events that often imply real-world consequences or richer information, we propose to exploit such commonsense attributes as “anchors”. It may ensure the learned representations are well associated via reasoning effects (Liu et al. 2020; Majumder et al. 2020) beyond the semantically close alignments.

Specifically, we populate an augmented dataset 𝒟~Link\tilde{\mbox{${\cal D}$}}_{Link} that expands pairs in 𝒟Link{\mbox{${\cal D}$}}_{Link} based on commonsense knowledge. For commonsense expansion, we capture implicit attributes of either personas and utterances and annotate them as metadata, using GPT2 based commonsense knowledge generators (Hwang et al. 2021) (Appendix A). The commonsense attributes are surrounded by special tokens and concatenated into a single sequence with persona or utterance.

 
Dialogue model Linking model R@1/20 \uparrow R@5/20 \uparrow MRR \uparrow Contradict@1 \downarrow
 
Bi-encoder N/A 0.814 0.973 0.882 0.075
Cross-encoder 0.884 0.991 0.930 0.040
Bi-encoder Manual Paraphrasing 0.758 0.961 0.876 0.088
Paraphrasing 0.769 0.963 0.851 0.094
Bi-encoder 0.881 0.987 0.927 0.053
Cross-encoder 0.881 0.988 0.928 0.043
Bi-encoder Persona-Link (small PKB) 0.904 0.992 0.943 0.039
Persona-Link (large PKB) 0.931 0.993 0.959 0.027
 
Table 1: The performance of dialogue models on Persona-Chat testset
Definition 2

(Commonsense Persona-Link Dataset) Given 𝒟Link={(ui,pi,yi)}i=1M\mbox{${\cal D}$}_{\textsc{Link}}=\{(u_{i},p_{i},y_{i})\}_{i=1}^{M}, its commonsense-expanded set is defined as 𝒟~Link={(u~i,p~i,yi)}i=1M\tilde{\mbox{${\cal D}$}}_{\textsc{Link}}=\{(\tilde{u}_{i},\tilde{p}_{i},y_{i})\}_{i=1}^{M} where expanded samples u~i\tilde{u}_{i} and p~i\tilde{p}_{i} are obtained by a commonsense expansion function ψ:(ui or pi)(u~i or p~i)\psi:(u_{i}\text{~{}or~{}}p_{i})\rightarrow(\tilde{u}_{i}\text{~{}or~{}}\tilde{p}_{i}). This function provides tuples that belong to nine relation types spanning over cause-effect interrelations between events: xAttr, xEffect, xIntent, xNeed, xReact, xWant, oEffect, oReact, oWant— where a prefix ‘x’ indicates an effect or cause on the person and ‘o’ denotes the same on others. We present more details of the commonsense expansion function in Appendix A.

Phase 3: Learning with Label Regularization

As many commonsense attributes are commonly shared between utterance and persona, they are likely to be relevant to each other although the NLI-driven label indicates negative. On the other hand, not always such an assumption is true since generated commonsense attributes are often ambiguous or over-claimed. These concerns motivate us to better regularize learning on 𝒟~Link\tilde{\mbox{${\cal D}$}}_{\textsc{Link}} by well-calibrated soft labels. That is, the expanded set 𝒟~Link\tilde{\mbox{${\cal D}$}}_{\textsc{Link}} on the denser space may not strictly follow the pre-annotated binary labels from NLI.

To address this, we argue that the linking model trained on 𝒟Link\mbox{${\cal D}$}_{\textsc{Link}} can be a good reference point. Specifically, we first train a linking model θLink\theta_{\textsc{Link}} only with the semantic-level Persona-Link dataset 𝒟Link\mbox{${\cal D}$}_{\textsc{Link}}, which can further perform inference over new data samples (u~,p~)(\tilde{u},\tilde{p}) to compute their outputs as new labels P(p~|u~;θLink)P(\tilde{p}|\tilde{u};\theta_{\textsc{Link}}). By regularizing by the semantic feature space of θLink\theta_{\textsc{Link}}, the commonsense-expanded dataset can have well-calibrated labels for encoding commonsense attributes. We thus train another linking model θ~Link\tilde{\theta}_{\textsc{Link}} with dual goals of following not only the original hard labels but also new soft labels as:

θLink=argminθ\displaystyle\theta_{\textsc{Link}}=\text{argmin}_{\theta} (u,p,y)𝒟Link(y,P(p|u;θ))\displaystyle\sum_{(u,p,y)\in\mathcal{D}_{\textsc{Link}}}\mbox{${\cal L}$}(y,P(p|u;\theta)) (3)
θ~Link=argminθ\displaystyle\tilde{\theta}_{\textsc{Link}}=\text{argmin}_{\theta} (u~,p~,y)𝒟~Link(y,P(p~|u~;θ))\displaystyle\sum_{(\tilde{u},\tilde{p},y)\in\tilde{\mathcal{D}}_{\textsc{Link}}}\mbox{${\cal L}$}(y,P(\tilde{p}|\tilde{u};\theta)) (4)
+λ(P(p~|u~;θLink),P(p~|u~;θ))\displaystyle+\lambda\cdot\mbox{${\cal L}$}(P(\tilde{p}|\tilde{u};\theta_{\textsc{Link}}),P(\tilde{p}|\tilde{u};\theta))

where λ\lambda is a preference weight with the distillation loss. To compute the linking score P(p|u;θ)P(p|u;\theta) and the cross-entropy loss {\cal L}, based on the Bi-encoder architecture allowing for fast and real-time inference, we follow the optimization procedure (Logeswaran et al. 2019; Humeau et al. 2020; Jeong et al. 2021) of retrieval-based models (e.g., information retrieval, entity linking, and response selection): The network is trained to maximize the score of the correct persona pp with respect to the (randomly sampled) personas of the same batch (i.e., in-batch negatives).

Inference

Once a linking model θ~Link\tilde{\theta}_{\textsc{Link}} is learned, given a utterance uu in Persona-Chat, we follow Eq. (2) to augment a new persona into the original dialogue episode. Considering that the development/test sets of Persona-Chat are unseen, as PKB \mathbb{P}, we only use a list of personas involved in the training set of Persona-Chat at least once.

 
Linking model Dialogue model
Pretrained on Wikipedia Pretrained on Reddit
R@1/20\uparrow R@5/20 \uparrow MRR \uparrow Ctrd.@1 \downarrow R@1/20\uparrow R@5/20 \uparrow MRR \uparrow Ctrd.@1 \downarrow
 
No persona (0%)
N/A 0.591 0.870 0.711 0.174 0.659 0.900 0.766 0.144
Bi-encoder 0.699 0.936 0.802 0.120 0.773 0.960 0.853 0.094
Persona-Link 0.733 0.956 0.828 0.113 0.805 0.970 0.876 0.078
Incomplete persona (80%)
N/A 0.768 0.953 0.848 0.094 0.825 0.973 0.889 0.073
Bi-encoder 0.768 0.966 0.850 0.094 0.845 0.984 0.905 0.064
Persona-Link 0.785 0.972 0.864 0.093 0.858 0.985 0.913 0.058
Original persona (100%)
N/A 0.814 0.973 0.882 0.075 0.868 0.986 0.919 0.054
 
Table 2: The response quality of Persona-Chat models, observing debiased personas added at test time. In advance, in-dialogue persona sentences were randomly removed, and link models augmented in-context personas based on dialogue histories.

Persona-Chat Evaluation

As bias analysis in benchmark datasets is a non-trivial problem (Bowman and Dahl 2021; Torralba and Efros 2011), we measure the response quality of dialogue models as a proxy of bias mitigation. We measure Recall@k/Nk/N and MRR as the model performance to the gold utterance, where N=20N=20 and k=[1,5]k=[1,5]. Another metric is Contradict@1, indicating the textual disagreement judged by an NLI model: the proportion of contradictory responses in the top-1 candidates returned by dialogue agents, i.e., consistency error ratio.

Performance with White-box Training

For the experiment on white-box settings (i.e., augmenting personas in both training and inference time), we consider following six models including two Persona-Link variants as the candidates to be analyzed.

1) Paraphrasing: As unsupervised data augmentation (Xie et al. 2020), we consider an off-the-shelf paraphrasing system in (Mallinson, Sennrich, and Lapata 2017), where personas were translated into a foreign language and back-translated as paraphrases.

2) Manual Paraphrasing: As labor-intensive augmentation, we use manually revised persona sentences additionally presented in Persona-Chat, where extra workers rephrased them to remove trivial word overlap.

3) Bi-encoder: As an IR baseline of linking model (Wu et al. 2020) trained on 𝒟Link\mbox{${\cal D}$}_{\textsc{Link}}, we use two transformers that separately encode utterance and persona, and multi-sentence scoring over the pre-computed cache of personas.

4) Cross-encoder: We consider using a single transformer (Humeau et al. 2020) that uses a richer self-attention mechanism, jointly encoding utterance and persona in 𝒟Link\mbox{${\cal D}$}_{\textsc{Link}} to obtain the final representation, which is impractical for real-time use.

5) Persona-Link (small PKB): Our proposed linking model with a smaller Persona Knowledge Base (PKB), caching 1K personas which are randomly sampled from the primal dataset 𝒟Chat\mathcal{D}_{\textsc{Chat}}.

6) Persona-Link (large PKB): Using the same model, we present the original version of our approach, where all 5K personas from 𝒟Chat\mathcal{D}_{\textsc{Chat}} are plugged into its PKB.

Table 1 shows the overall performance of dialogue models trained with different linking models. Models paired with Persona-Link variants significantly outperform other baselines for all measures. This suggests that training dialogue model with well augmented personas helps the model to be more persona-grounded and consistent which is a primary goal in persona-based dialogue literature. Furthermore, as the performance gain increases with the size of PKB used for Persona-Link, we believe that populating bigger PKB improves the performance. Another observation is that Persona-Link variants outperform both manual and automatic paraphrasing approaches on every measure. Even though paraphrasing techniques might contribute to richer representation of personas, we argue that the paraphrasing approaches are limited to mitigating the trivial word overlap while utterances can be expressed in much flexible way in persona-based dialogue. While Bi- and Cross-encoder linking models result in better performance compare to the paraphrasing-based linking models due to out-dialogue matching which aims to mitigate linguistic bias, Persona-Link variants outperform in both measures. Especially, we observe all Persona-Link variants outperform Cross-encoder which compromises high computational cost for performance. We further analyze how Persona-Link achieves better performance in Persona-Link Evaluation section.

Performance with Black-box Training

To demonstrate that the augmented persona is not over-claimed by out-dialogue matching or out-corpus expansion, we compare the quality of augmented persona using dual task models. For that, we simulate a setting where none or few persona information is provided and new personas are augmented only in inference time by using off-the-shelf dialogue agents (i.e., learned without augmenting personas in training time). Inspired by the evaluation method of measuring the relevance of generated response in (Su et al. 2020), we use the pre-trained dialogue agents as our diagnosis model assuming that the response quality of the model reflects the suitability of augmented persona in given dialogue contexts. That is, the suitability of the augmented persona to the context would be low if the augmented personas, which replace partial or entire persona of dataset, cause the failure of pre-trained dialogue agents. For the diagnosis agents, we adopt two dialogue models of Bi-encoder; each pretrained on Wikipedia (BERT-like; Devlin et al. 2019) or Reddit (an adapted setup for dialogue; Mazaré et al. 2018), and both fine-tuned on 𝒟Chat\mbox{${\cal D}$}_{\textsc{Chat}}. Note that, during fine-tuning, the dialogue agents only observe the original personas. As dual task models, we adopt two linking models; Bi-encoder trained on 𝒟Link\mbox{${\cal D}$}_{\textsc{Link}} and our Persona-Link trained on 𝒟~Link\tilde{\mbox{${\cal D}$}}_{\textsc{Link}}.

Specifically, the original persona sentences were removed at a given percentage; No persona eradicates the whole persona information and Incomplete persona discards 20%. In test time, linking models generate personas based on the utterance histories and merge them into the sample for testing dialogue models222For fair comparison, all utterances on the development set remained unseen by linking models. Also, for faster computation, all utterances in dialogue were concurrently processed in batch.. With persona augmentation, Table 2 shows the overall performance of dialogue models. In each setting, we report the base performance with information limit, where no linking model is available, i.e., N/A. In the last row, we report the upper bound performance, where dialogue models access full information of Original persona.

First, compared to N/A, we observe linking models significantly improve (+14.6% at most) response quality by backtracking personas from utterances for the augmentation. Dialogue models seem to benefit from augmented personas, even though the models haven’t seen them in training process, suggesting that personas provided by Persona-Link successfully keep contextual relevance to the given dialogues. Second, Persona-Link models outperform Bi-encoder models in all settings. As the only difference between the two is commonsense expansion with regularized training, our approach seems to enrich dual task with common knowledge beyond the alignments in Persona-Chat. Thirdly, the performance gain is the greatest in the persona-free setup, which is common in real-world scenarios. This suggests that our persona augmentation may boost dialogue agents in bootstrapping user profiles on-the-fly. As observing only debiased personas improves a trained agent, we conclude that our approach efficiently counteracts linguistic bias in Persona-Chat.

Now, we examine how persona augmentation calibrates model prediction in response retrieval, by showing an example in Table 3. After removing a persona in advance, a linking model augments its own personas into context based on dialogue history. Observing additional information relevant to utterances (i.e.meal) and linguistically debiased (i.e.steak), the model prediction was calibrated online to retrieve the gold response, which is more consistent and engaging. Note that we use the same dialogue model to retrieve both responses. (See Appendix C for details).

 
Agent’s personas:
My mom is a secretary.
I am a bodybuilder.
I have one brother.
(-) I like to eat a lot of meat.
(+) I am a meat eater.
(+) My favorite meal is steak.
 
Dialogue history:
User: Hi there, how are you tonight?
Agent: Great. I just finished a huge steak. How are you?
User: I am good, drinking some scotch.
Response without augmented personas:
Agent: Cool! What are your hobbies?
Response with augmented personas:
Agent: I am major meat eater to build muscles.
 
Table 3: The model calibration in response retrieval task based on online persona augmentation. After removing persona (-), a linking model augmented personas (+), which calibrates the prediction of dialogue agent, with any parameter updates.

Persona-Link Evaluation

In this section, we conduct additional experiments to evaluate the proposed process for the dual task.

Overall Linking Performance

To study the model accuracy to the dual task, we perform comparative experiments of Persona-Link. To find ground truth pairs for Persona-Link, we manually collect 300 utterance-persona pairs from the test data in Persona-Chat, including 230 unique ground truth personas with seven annotators, which achieve a very good agreement (Cohen’s kappa = 0.86). The number of utterances per persona is 1.16. As in Persona-Chat evaluation, we measure Recall@kk and MRR as the model performance to the gold persona.

We consider five IR baselines trained on 𝒟Link\mathcal{D}_{\textsc{Link}}: Cosine Similarity, BM25 (Robertson et al. 1995), K-NRM (Xiong et al. 2017) Conv-KNRM (Dai et al. 2018), and Bi-encoder (Humeau et al. 2020). Also, to investigate the impact of commonsense expansion and label regularization, we consider two variants of our approach using alternative expansion methods: EDA (Wei and Zou 2019), and Paraphrasing (Mallinson, Sennrich, and Lapata 2017). More details for these baselines are presented in Appendix B.

 
Linking Model R@1 R@10 MRR
 
Cosine Similarity 0.108 0.349 0.190
BM25 0.491 0.792 0.595
K-NRM 0.294 0.591 0.384
CONV-KNRM 0.345 0.628 0.439
Bi-encoder 0.583 0.871 0.684
Ours (EDA) 0.599 0.885 0.709
Ours (Para) 0.610 0.885 0.715
Ours 0.669 0.922 0.759
 
Table 4: Linking performance on Persona-Link
Refer to caption
(a) w/o commonsense expansion
Refer to caption
(b) w/ commonsense expansion
Figure 3: t-SNE visualization on the development set. The colored points indicate the utterance embeddings relevant to persona “I have a marketing job”.
Refer to caption
Figure 4: Similarity with respect to linking method.

Table 4 shows the overall linking performance. First, we observe that among IR baselines trained on 𝒟Link\mathcal{D}_{\textsc{Link}}, Bi-encoder achieves the best performance, which validates our choice of base model for the dual task. Second, all models trained on 𝒟Link\mathcal{D}_{\textsc{Link}} show poorer results than Ours and its variants using expansion methods with regularized training. This means that leveraging expanded dataset and calibrating labels into soft labels is more appropriate than naive IR models. Finally, Ours outperforms its variants (and other linking models as well) in all metrics. We note that while utterance or persona remains semantically similar by EDA or paraphrasing, Ours expands some commonsense attributes so that potentially relevant mapping can be captured. This supports the efficacy of exploiting commonsense as anchors, which may associates learned representations beyond the semantically close alignments. In Figure 3, we visualize the utterance embeddings without (Figure 3a) and with (Figure 3b) commonsense expansion. We observe that the utterance embeddings relevant to a persona gets closer with commonsense expansion, which can be a possible explanation for the performance boost of Ours exploiting commonsense over its variants.

Refer to caption
Figure 5: Performance with respect to PI result.
Refer to caption
Figure 6: Performance with respect to data size.

Analysis on Linguistic Bias

In addition to the linking performance, we further analyze the linked utterance-to-persona alignments by linking models to see their linguistic bias. We plot the linguistic similarity between utterance-to-persona alignments in Figure 4. Following (Park et al. 2019), we calculate Jaccard similarity and METEOR, widely known metric for lexical similarity and semantic similarity, respectively. Jaccard similarity measures word co-occurrence and METEOR measures exact, stem, synonym, and paraphrase matches. For both metrics, an alignment with lower similarity can be seen as the one with less linguistic bias. Two models drop the original similarity from 0.2 to 0.1 on average, indicating the efficacy of out-dialogue semantic matching. Ours shows a slight decrease compared to out-dialogue model which confirms that Ours successfully leverages the commonsense expansion to align utterances to debiased persona.

We compare the performance of θLink\theta_{\textsc{Link}} and θ~Link\tilde{\theta}_{\textsc{Link}} in semantically different pair sets to investigate whether commonsense expansion helps the model link to debiased persona. Here, we use a BERT model fine-tuned with a popular paraphrase identification (PI) dataset called Quora Question Pairs to split our utterance-persona pairs into semantically equivalent and non-equivalent sets. With this model, only 35.5% of utterance-persona pairs are inferred as positive (i.e., paraphrase). For analysis, in Figure 5, we split the overall utterance-persona pairs into positive and negative sets by the PI model, and ablate the performance of linking model with or without using commonsense knowledge and label regularization. As a result, while θ~Link\tilde{\theta}_{\textsc{Link}} shows better performance than θLink\theta_{\textsc{Link}} in all sets, θ~Link\tilde{\theta}_{\textsc{Link}} achieves much better performance in negative set of R@10. This demonstrates the advantage of using commonsense knowledge, which gives the model the ability to link semantically different pairs that leads θ~Link\tilde{\theta}_{\textsc{Link}} to link utterance to debiased persona.

Next, we evaluate how commonsense expansion affects when only a limited number of examples are provided in train time. Figure 6 shows the linking performance grouped by frequency of utterance-persona pairs in the training data in terms of R@10. Here, the performance means how well the model links utterances to their gold persona when the utterance-persona pair of the gold persona is not provided or few in train time. We observe that θ~Link\tilde{\theta}_{\textsc{Link}} has more balanced performance for any persona groups, which is the effect of using commonsense knowledge, which can work as pivots when the utterance and persona has similar commonsense in common. Especially, θ~Link\tilde{\theta}_{\textsc{Link}} shows the significant improvement for unseen personas (w/o any aligned training data) compared to θLink\theta_{\textsc{Link}}, which shows that θ~Link\tilde{\theta}_{\textsc{Link}} may be applicable to zero-shot linking. Since θ~Link\tilde{\theta}_{\textsc{Link}} shows great performance to unseen utterance-persona pairs, we suggest to use our linking model with a large set of personas.

Related Work

Building personalized dialogue agents has been a popular task recently. Zhang et al. (2018) introduced the Persona-Chat dataset, a crowd-sourced conversation dataset with persona information, to improve model engagingness and consistency. Based on such data, recent works focus on improving persona-grounded dialogue agent’s performance (Kim, Kim, and Kim 2020; Li et al. 2020; Song et al. 2020; Huang, Zhu, and Gao 2020; Zhang et al. 2019; Luo et al. 2019). Yavuz et al. (2019) designed the DeepCopy model, which leverages copy mechanism to incorporate persona texts. Song et al. (2019) integrated persona texts into the Per-CVAE model for generating diverse responses. However, prior work reports several linguistic bias which hamper learning to ground more engaging and consistent utterances.

Few recent works focused on augmenting grounding with commonsense knowledge with successful applications in open-domain dialogue agent (Ghazvininejad et al. 2018; Moon et al. 2019). Ghazvininejad et al. (2018) generalizes the Seq2Seq approach to neural conversation models by combining conversational and KG data. Moon et al. (2019) propose DialKG walker model that learns the transitions of dialogue contexts as structured traversals over KG. In this work, we extend this effort into improving persona-grounded dialogue dataset in data-centric view via augmenting relevant personas.

Conclusion

In this work, we propose a primal-dual task framework to de-bias the Persona-Chat dataset and its dialogue model without any human effort. By exploiting external LMs and knowledge distillation in a systematic way of overcoming the linguistic bias in Persona-Chat, our linking model was effective in improving 11.7% in dialogue accuracy from BERT Bi-encoder model using the original Persona-Chat dataset. We hope future research to leverage the graph structure of PKB with commonsense attributes for better persona linking.

Acknowledgements

This work was partly supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (No. 2020-0-01361, Artificial Intelligence Graduate School Program (Yonsei University)) and the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2020-11-0863). Jinyoung Yeo is a corresponding author.

References

  • Bosselut et al. (2019) Bosselut, A.; Rashkin, H.; Sap, M.; Malaviya, C.; Celikyilmaz, A.; and Choi, Y. 2019. COMET: Commonsense transformers for automatic knowledge graph construction. In Proceedings of ACL.
  • Bowman and Dahl (2021) Bowman, S.; and Dahl, G. 2021. What Will it Take to Fix Benchmarking in Natural Language Understanding? In Proceedings of ACL.
  • Dai et al. (2018) Dai, Z.; Xiong, C.; Callan, J.; and Liu, Z. 2018. Convolutional neural networks for soft-matching n-grams in ad-hoc search. In Proceedings of WSDM.
  • Devlin et al. (2019) Devlin, J.; Chang, M.-W.; Lee, K.; and Toutanova, K. 2019. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of NAACL-HLT.
  • Ghazarian et al. (2021) Ghazarian, S.; Liu, Z.; Chakrabarty, T.; Ma, X.; Galstyan, A.; and Peng, N. 2021. DiSCoL: Toward Engaging Dialogue Systems through Conversational Line Guided Response Generation. arXiv preprint arXiv:2102.02191.
  • Ghazvininejad et al. (2018) Ghazvininejad, M.; Brockett, C.; Chang, M.-W.; Dolan, B.; Gao, J.; Yih, W.-t.; and Galley, M. 2018. A knowledge-grounded neural conversation model. In Proceedings of AAAI.
  • Huang, Zhu, and Gao (2020) Huang, M.; Zhu, X.; and Gao, J. 2020. Challenges in building intelligent open-domain dialog systems. ACM Transactions on Information Systems.
  • Humeau et al. (2020) Humeau, S.; Shuster, K.; Lachaux, M.-A.; and Weston, J. 2020. Poly-encoders: Transformer architectures and pre-training strategies for fast and accurate multi-sentence scoring. In Proceedings of ICLR.
  • Hwang et al. (2021) Hwang, J. D.; Bhagavatula, C.; Le Bras, R.; Da, J.; Sakaguchi, K.; Bosselut, A.; and Choi, Y. 2021. (Comet-) Atomic 2020: On Symbolic and Neural Commonsense Knowledge Graphs. In Proceedings of AAAI.
  • Jeong et al. (2021) Jeong, M.; Choi, S.; Yeo, J.; and Hwang, S.-w. 2021. Label and Context Augmentation for Response Selection at DSTC8. IEEE/ACM Transactions on Audio, Speech, and Language Processing.
  • Kim, Kim, and Kim (2020) Kim, H.; Kim, B.; and Kim, G. 2020. Will I Sound like Me? Improving Persona Consistency in Dialogues through Pragmatic Self-Consciousness. In Proceedings of EMNLP.
  • Levy and Goldberg (2014) Levy, O.; and Goldberg, Y. 2014. Dependency-based word embeddings. In Proceedings of ACL.
  • Li et al. (2020) Li, A. W.; Jiang, V.; Feng, S. Y.; Sprague, J.; Zhou, W.; and Hoey, J. 2020. ALOHA: Artificial Learning of Human Attributes for Dialogue Agents. In Proceedings of AAAI.
  • Liu et al. (2019) Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; and Stoyanov, V. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint.
  • Liu et al. (2020) Liu, Y.; Yang, T.; You, Z.; Fan, W.; and Yu, P. S. 2020. Commonsense Evidence Generation and Injection in Reading Comprehension. In Proceedings of SIGDIAL.
  • Logeswaran et al. (2019) Logeswaran, L.; Chang, M.-W.; Lee, K.; Toutanova, K.; Devlin, J.; and Lee, H. 2019. Zero-shot entity linking by reading entity descriptions. In Proceedings of ACL.
  • Luo et al. (2019) Luo, L.; Huang, W.; Zeng, Q.; Nie, Z.; and Sun, X. 2019. Learning personalized end-to-end goal-oriented dialog. In Proceedings of AAAI.
  • Majumder et al. (2020) Majumder, B. P.; Jhamtani, H.; Berg-Kirkpatrick, T.; and McAuley, J. 2020. Like hiking? You probably enjoy nature: Persona-grounded Dialog with Commonsense Expansions. In Proceedings of EMNLP.
  • Mallinson, Sennrich, and Lapata (2017) Mallinson, J.; Sennrich, R.; and Lapata, M. 2017. Paraphrasing revisited with neural machine translation. In Proceedings of EACL.
  • Mazaré et al. (2018) Mazaré, P.-E.; Humeau, S.; Raison, M.; and Bordes, A. 2018. Training millions of personalized dialogue agents. In Proceedings of EMNLP.
  • Moon et al. (2019) Moon, S.; Shah, P.; Kumar, A.; and Subba, R. 2019. Opendialkg: Explainable conversational reasoning with attention-based walks over knowledge graphs. In Proceedings of ACL.
  • Park et al. (2019) Park, S.; Hwang, S.-w.; Chen, F.; Choo, J.; Ha, J.-W.; Kim, S.; and Yim, J. 2019. Paraphrase diversification using counterfactual debiasing. In Proceedings of AAAI.
  • Radford et al. (2018) Radford, A.; Narasimhan, K.; Salimans, T.; and Sutskever, I. 2018. Improving language understanding by generative pre-training. OpenAI Blog.
  • Radford et al. (2019) Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; and Sutskever, I. 2019. Language Models are Unsupervised Multitask Learners.
  • Robertson et al. (1995) Robertson, S. E.; Walker, S.; Jones, S.; Hancock-Beaulieu, M. M.; Gatford, M.; et al. 1995. Okapi at TREC-3. Nist Special Publication Sp.
  • Sap et al. (2019) Sap, M.; Le Bras, R.; Allaway, E.; Bhagavatula, C.; Lourie, N.; Rashkin, H.; Roof, B.; Smith, N. A.; and Choi, Y. 2019. Atomic: An atlas of machine commonsense for if-then reasoning. In Proceedings of AAAI.
  • Song et al. (2019) Song, H.; Zhang, W.-N.; Cui, Y.; Wang, D.; and Liu, T. 2019. Exploiting persona information for diverse generation of conversational responses. In Proceedings of IJCAI.
  • Song et al. (2020) Song, H.; Zhang, W.-N.; Hu, J.; and Liu, T. 2020. Generating persona consistent dialogues by exploiting natural language inference. In Proceedings of AAAI.
  • Su et al. (2020) Su, H.; Shen, X.; Zhao, S.; Xiao, Z.; Hu, P.; Zhong, R.; Niu, C.; and Zhou, J. 2020. Diversifying Dialogue Generation with Non-Conversational Text. In Proceedings of ACL.
  • Su, Chuang, and Chen (2020) Su, S.-Y.; Chuang, Y.-S.; and Chen, Y.-N. 2020. Dual Inference for Improving Language Understanding and Generation. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings, 4930–4936.
  • Torralba and Efros (2011) Torralba, A.; and Efros, A. A. 2011. Unbiased look at dataset bias. In CVPR 2011, 1521–1528. IEEE.
  • Wang et al. (2019) Wang, A.; Singh, A.; Michael, J.; Hill, F.; Levy, O.; and Bowman, S. R. 2019. Glue: A multi-task benchmark and analysis platform for natural language understanding. In Proceedings of ICLR.
  • Wang et al. (2021) Wang, J.; Chen, K.; Shou, L.; Wu, S.; and Chen, G. 2021. Effective Slot Filling via Weakly-Supervised Dual-Model Learning. In Proceedings of AAAI.
  • Wei and Zou (2019) Wei, J.; and Zou, K. 2019. EDA: Easy data augmentation techniques for boosting performance on text classification tasks. In Proceedings of EMNLP-IJCNLP.
  • Welleck et al. (2018) Welleck, S.; Weston, J.; Szlam, A.; and Cho, K. 2018. Dialogue natural language inference. In Proceedings of ACL.
  • Wolf et al. (2019) Wolf, T.; Sanh, V.; Chaumond, J.; and Delangue, C. 2019. Transfertransfo: A transfer learning approach for neural network based conversational agents. arXiv preprint arXiv:1901.08149.
  • Wu et al. (2020) Wu, L.; Petroni, F.; Josifoski, M.; Riedel, S.; and Zettlemoyer, L. 2020. Scalable Zero-shot Entity Linking with Dense Entity Retrieval. In Proceedings of EMNLP.
  • Xia et al. (2017) Xia, Y.; Qin, T.; Chen, W.; Bian, J.; Yu, N.; and Liu, T.-Y. 2017. Dual supervised learning. In International Conference on Machine Learning, 3789–3798. PMLR.
  • Xie et al. (2020) Xie, Q.; Dai, Z.; Hovy, E.; Luong, M.-T.; and Le, Q. V. 2020. Unsupervised data augmentation for consistency training.
  • Xiong et al. (2017) Xiong, C.; Dai, Z.; Callan, J.; Liu, Z.; and Power, R. 2017. End-to-End Neural Ad-hoc Ranking with Kernel Pooling. In Proceedings of SIGIR.
  • Yavuz et al. (2019) Yavuz, S.; Rastogi, A.; Chao, G.-L.; and Hakkani-Tur, D. 2019. DeepCopy: Grounded Response Generation with Hierarchical Pointer Networks. In Proceedings of the 20th Annual SIGdial Meeting on Discourse and Dialogue, 122–132.
  • Zhang et al. (2018) Zhang, S.; Dinan, E.; Urbanek, J.; Szlam, A.; Kiela, D.; and Weston, J. 2018. Personalizing dialogue agents: I have a dog, do you have pets too? In Proceedings of ACL.
  • Zhang et al. (2019) Zhang, Y.; Gao, X.; Lee, S.; Brockett, C.; Galley, M.; Gao, J.; and Dolan, B. 2019. Consistent dialogue generation with self-supervised feature learning. arXiv preprint.
  • Zhu, Cao, and Yu (2020) Zhu, S.; Cao, R.; and Yu, K. 2020. Dual learning for semi-supervised natural language understanding. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 28: 1936–1947.

Appendix A Appendix A: Implementation Details

COMET

We use COMET (Bosselut et al. 2019) which generates rich and diverse commonsense expansions of a given world event.

It is a fine-tuned version of a pre-trained GPT2 (Radford et al. 2018) model on a pre-existing commonsense knowledge graph such as ATOMIC (Sap et al. 2019) that can generate novel nodes (events) and edges (relations). Specifically, ATOMIC provides tuples that belong to nine relation types spanning over cause-effect interrelations between events: xAttr, xEffect, xIntent, xNeed, xReact, xWant, oEffect, oReact, oWant— where a prefix ‘x’ indicates an effect or cause on the person and ‘o’ denotes the same on others. While we tried COMET finetuned on commonsense other than ATOMIC, such as ConceptNet, we found expansions unsatisfactory and do not report them.333Failure case includes that event-like persona (‘I joined the air force, … I learned how to fly!’) fails to match concept ‘fly’. For more details on COMET and ATOMIC, we refer the reader to (Bosselut et al. 2019) and (Sap et al. 2019) respectively.

We use the COMET framework to generate expansions for not only persona but also utterance along the multiple relation types that ATOMIC provides (Figure 2). Here we use only 6 ATOMIC relations with the prefix ‘x’, since Persona-Link targets the utterances that the agent is talking about itself. We obtain different samples while decoding via greedy search from COMET for more accurate expansions (sacrificing diversity compared to beam search).

For COMET expansions, we use the code444https://github.com/atcbosselut/comet-commonsense released by the authors of COMET. We follow the default setting for this code to generate commonsense knowledge.

MNLI model

As a MNLI model, we use the public HuggingFace implementation of the RoBERTa-large model (Liu et al. 2019) fine-tuned with the Multi-Genre NLI dataset (Wang et al. 2019).

 
Personas
I like to remodel homes.
I like to go hunting.
I like to shoot a bow.
My favorite holiday is Halloween.
 
Dialogue history
Person1: Hi, how are you doing? I am getting ready to
                  do some cheetah chasing to stay in shape.
Person2: You must be very fast.
                  Hunting is one of my favorite hobbies.
Person1: I am! For my hobby
                  I like to do canning or some whittling.
Person2: I also remodel homes
                  when I am not out bow hunting.
Person1: That is neat. When I was in high school
                  I placed 6th in 100m dash!
Person2: That is awesome.
                  Do you have a favorite season or time of year?
Person1: I do not. But I do have a favorite meat
                  since that is all I eat exclusively.
Person2: What is your favorite meat to eat?
Person1: I would have to say its prime rib.
                  Do you have any favorite foods?
Person2: I like chicken or macaroni and cheese.
Person1: Do you have anything planned for today?
                  I think I am going to do some canning.
Person2: I am going to watch football.
                  What are you canning?
Person1: I think I will can some jam.
                  Do you also play footfall for fun?
Person2: If I have time outside of hunting
                  and remodeling homes.
                  Which is not much!
 
Table 5: An example of Persona-Chat dataset
 
Utterance Persona
I do not. I spend all of my time playing frisbee. I like playing frisbee.
Terrific. My favorite thing to do is dance. I like dance.
I love having hair on my face. I have a beard.
You must like pizza as much as I like rush. You know? The band? I love the band rush.
I am a nurse. I work in the er and surgery unit. I work at the hospital.
I do actually most performance art tho. Some abstract. I am an artist.
I have a pet fish. I have an exotic fish aquarium.
 
Table 6: Examples of Persona-Link prediction
 
Agent’s personas:
My mom is a secretary.
I am a bodybuilder.
I have one brother.
(-) I like to eat a lot of meat.
(+) I am a meat eater.
(+) My favorite meal is steak.
 
Dialogue history:
User: Hi there, how are you tonight?
Agent: Great. I just finished a huge steak. How are you?
User: I am good, drinking some scotch.
Response without augmented personas:
Agent: Cool! What are your hobbies?
Response with augmented personas:
Agent: I am major meat eater to build muscles.
User: Oh, I like meat ok.
Agent: Scotch goes good with meat.
User: Scotch goes good with anything lol.
Agent: I have a brother who likes scotch.
User: I am an only child.
Agent: I have just the one brother and 3 sisters.
User: Do you have any pets? I am scared of dogs.
Agent: I have a cat and a dog.
You would not be scared of my dog.
User: What do you do for a living?
Agent: I am a secretary like my mom hard living and you?
User: I am an author. I write short stories.
Agent: Do you want to write a story
             about a major meat eating bodybuilder?
 
Table 7: Full dialogue of Table 3

Persona-Link model

As a Persona-Link model, we use BERT-small as a transformer for efficiency. We use the Adam optimizer with a learning rate of 5e-5 with batch size 100, a max number of tokens of 64 for mention/persona encoder, epoch 10, weight decay 0.01, adam epsilon 1e-8, gradients clipping to 1, and distillation preference weight rate 1.0. The best parameters were chosen based on the Recall@10 score. Similar to Mazaré et al. (2018), during training we consider the other labels in the batch as negatives.

Persona-Chat model

We use the ParlAI implementation of Bi-encoder pre-trained on Reddit and fine-tuned on Persona-Chat as a dialogue agent. It is notable that the Bi-encoder for dialogue agent is different from that of Persona-Link model. While Persona-Link model links utterances to their referent personas, the dialogue agent predicts the next utterances in dialogue.

Bi-encoder

As aforementioned, we use Bi-encoder as a model architecture of both Persona-Chat and Persona-Link. Here, we describe the architecture of Bi-encoder and its scoring methods. First, each context cic_{i} and corresponding kk-th response candidate ri,kr_{i,k} are encoded with transformers TcontextT_{context} and TresponseT_{response} respectively:

yci=red(Tcontext(ci))\displaystyle y_{c_{i}}=red(T_{context}(c_{i}))
yri,k=red(Tresponse(ri,k))\displaystyle y_{r_{i,k}}=red(T_{response}(r_{i,k}))

where red()red(\cdot) is a function that reduces the sequence of vectors into a vector. The output of [CLS] token is used as red()red(\cdot) following (Humeau et al. 2020). Then, the score of each response is computed by the dot-product:

s(ci,ri,k)=yciyri,ks(c_{i},r_{i,k})=y_{c_{i}}\cdot y_{r_{i,k}}

In training time, the model is optimized by minimizing a cross-entropy loss. We use other correct responses in the same batch as negatives following (Humeau et al. 2020) on training.

Appendix B Appendix B: Baselines of Persona-Link Task

  • Cosine Similarity: This baseline selects a persona with the highest cosine similarity between the utterance and persona vectors using the pre-trained word2vecf (Levy and Goldberg 2014) not word counts.

  • BM25 (Robertson et al. 1995): A strong unsupervised retrieval model which is one of the most popular ranking functions in off-the-shelf search engines.

  • K-NRM (Xiong et al. 2017): A kernel-based efficient neural retrieval model that can consider both word-level soft/hard matching signals, i.e., semantic similarity via term embeddings and exact matches of query terms.

  • Conv-KNRM (Dai et al. 2018): A convolutional kernel-based neural ranking model that leverages n-gram soft matches.

  • Bi-encoder (Humeau et al. 2020): A state-of-the-art of transformer architecture for IR. We use BERT (Devlin et al. 2019) as LM.

  • Paraphrasing (Mallinson, Sennrich, and Lapata 2017): Utterances and personas are first translated to a foreign pivot, then then back-translated to English, producing a paraphrase as a new sample.

  • EDA (Wei and Zou 2019): A powerful data augmentation tool for text classification with synonym replacement, random insertion, random swap, and random deletion.

Appendix C Appendix C: Examples

An Example of Persona-Chat Dataset

Table 5 shows one of the dialogue in the Persona-Chat dataset. As shown, there are 4 persona sentences of Person2, which Person2 speaker imitates. For example, in the first utterance of Person2, Person2 says “Hunting is one of my favorite hobbies.” imitating the persona “I like to go hunting.”.

Examples of Persona-Link Prediction

Table 6 shows the predictions by Persona-Link model. Once trained, Persona-Link model links each utterance to its referent persona. For example, given “I am a nurse. I work in the er and surgery unit.” as an utterance, Persona-Link model predicts “I work at the hospital.” as a persona, which can be seen as appropriate and can be used for learning better persona-grounded dialogue agents.

Full Dialogue of Table 3

Table 7 in this section shows a full dialogue of Table 3 with the top ranked response by the dialog agent. Original personas, removed persona (-), augmented personas (+) by a linking model, dialogue history, and two responses without and with augmented personas are given in each slot.

That is, Table 7 presents retrieved response samples with comparison between without and with augmented personas. While a persona “I like to eat a lot of meat.” is removed, Persona-Link model can augment additional personas, such as “I am a meat eater.”, which supplement missing information. As a result, although the inferred personas are a little bit different with the original one, the dialogue agent model with them correctly responds to the user’s question, but the response without augmented personas is not the case.