SAPPHIRE: Approaches for Enhanced Concept-to-Text Generation
Abstract
We motivate and propose a suite of simple but effective improvements for concept-to-text generation called SAPPHIRE: Set Augmentation and Post-hoc PHrase Infilling and REcombination. We demonstrate their effectiveness on generative commonsense reasoning, a.k.a. the CommonGen task, through experiments using both BART and T5 models. Through extensive automatic and human evaluation, we show that SAPPHIRE noticeably improves model performance. An in-depth qualitative analysis illustrates that SAPPHIRE effectively addresses many issues of the baseline model generations, including lack of commonsense, insufficient specificity, and poor fluency.
1 Introduction
There has been increasing interest in constrained text generation tasks which involve constructing natural language outputs under certain pre-conditions, such as particular words that must appear in the output sentences. A related area of work is data-to-text natural language generation (NLG), which requires generating natural language descriptions of structured or semi-structured data inputs. Many constrained text generation and NLG tasks share commonalities, one of which is their task formulation: a set of inputs must be converted into natural language sentences. This set of inputs can be, in many cases, thought of as concepts, e.g. higher-level words or structures that play an important role in the generated text.
With the increased popularity of Transformer-based models and their application to many NLP tasks, performance on many text generation tasks has improved considerably. Much progress in recent years has been from the investigation of model improvements, such as larger and more effectively pretrained language generation models. However, are there simple and effective approaches to improving performance on these tasks that can come from the data itself? Further, can we potentially use the outputs of these models themselves to further improve their task performance - a “self-introspection” of sorts?
In this paper, we show that the answer is yes. We propose a suite of simple but effective improvements for concept-to-text generation called SAPPHIRE: Set Augmentation and Post-hoc PHrase Infilling and REcombination. Specifically, SAPPHIRE is composed of two major approaches: 1) the augmentation of input concept sets (§4.1), 2) the recombination of phrases extracted from baseline generations into more fluent and logical text (§4.2). These are mainly model-agnostic improvements that rely on the data itself and the model’s own initial generations, respectively.111Code at https://github.com/styfeng/SAPPHIRE
We focus on generative commonsense reasoning, or CommonGen Lin et al. (2020), which involves generating logical sentences describing an everyday scenario from a set of concepts, which in this case are individual words that must be represented in the output text in some form. CommonGen is a challenging instance of constrained text generation that assesses 1) relational reasoning abilities using commonsense knowledge, and 2) compositional generalization capabilities to piece together concept combinations. Further, CommonGen’s task formulation and evaluation methodology are quite broadly applicable and encompassing, making it a good benchmark for general constrained text generation capability. Further, this is an opportune moment to investigate this task as commonsense ability of NLP models, particularly for generation, has received increasing community attention through works like COMET Bosselut et al. (2019).
We perform experiments on varying sizes of two state-of-the-art Transformer-based language generation models: BART Lewis et al. (2020) and T5 Raffel et al. (2020). We first conduct an extensive correlation study (§3.1) and qualitative analysis (§3.2) of these models’ generations after simply training on CommonGen. We find that performance is positively correlated with concept set size, motivating concept set augmentation. We also find that generations contain issues related to commonsense and fluency which can possibly be addressed through piecing the texts back together in different ways, motivating phrase recombination.
Fleshing out our first intuition - we devise two methods to augment concepts from references during training through extracted keywords (§4.1.1) and attention matrices (§4.1.2). For the phrase recombination intuition, we propose two realizations based on a new training stage (§4.2.1) and masked infilling (§4.2.2). Finally, through comprehensive evaluation (§6), we show how the SAPPHIRE suite drives up model performance across metrics, besides addressing aforementioned baseline deficiencies on commonsense, specificity, and fluency.
2 Dataset, Models, and Metrics
2.1 CommonGen Dataset
The CommonGen dataset is split into train, dev, and test splits, covering a total of 35,141 concept sets and 79,051 sentences. The concept sets range from 3 keywords to 5 keywords long. As the original test set is hidden, we split the provided dev set into a new dev and test split for the majority of our experiments while keeping the training split untouched. Note that we also evaluate our SAPPHIRE models on the original test set with help from the CommonGen authors (see §6.1). We will henceforth refer to these new splits as trainCG, devCG, and testCG, and the original dev and test splits as devO and testO. The statistics of our new splits compared to the originals can be found in Table 1. We attempt to keep the relative sizes of the new dev and test splits and the distribution of concept set sizes within each split similar to the originals.
Dataset Stats | TrainCG | DevO | TestO | DevCG | TestCG |
# concept sets | 32,651 | 993 | 1,497 | 240 | 360 |
size = 3 | 25,020 | 493 | - | 120 | - |
size = 4 | 4,240 | 250 | 747 | 60 | 180 |
size = 5 | 3,391 | 250 | 750 | 60 | 180 |
# sentences | 67,389 | 4,018 | 7,644 | 984 | 1583 |
2.2 Models: T5 and BART
We perform experiments using pretrained language generators, specifically BART and T5 (both base and large versions). BART Lewis et al. (2020) is a Transformer-based seq2seq model trained as a denoising autoencoder to reconstruct original text from noised text. T5 Raffel et al. (2020) is another seq2seq Transformer with strong multitask pretraining. We use their HuggingFace codebases.
We train two seeded instances of each model on trainCG, evaluating their performance on devO, and comparing our numbers to those reported in Lin et al. (2020) to benchmark our implementations. These essentially serve as the four baseline models for our ensuing experiments. We follow the hyperparameters from Lin et al. (2020), choose the epoch reaching highest ROUGE-2 on the dev split, and use beam search for decoding.222See Appendix A for further details. From Table 2, we see that our re-implemented models match or exceed the original reported results on most metrics across different models.
Model\Metrics | BLEU-4 | CIDEr | SPICE |
Reported BART-large | 27.50 | 14.12 | 30.00 |
Reported T5-base | 18.00 | 9.73 | 23.40 |
Reported T5-Large | 30.60 | 15.84 | 31.80 |
Our BART-base | 28.30 | 15.07 | 30.35 |
Our BART-large | 30.20 | 15.72 | 31.20 |
Our T5-base | 31.00 | 16.37 | 32.05 |
Our T5-large | 33.60 | 17.02 | 33.45 |
2.3 Evaluation Metrics
For our experiments, we use a gamut of automatic evaluation metrics. These include those used by Lin et al. (2020), such as BLEU Papineni et al. (2002), CIDEr Vedantam et al. (2015), SPICE Anderson et al. (2016), and Coverage (Cov). Barring Cov, these metrics measure the similarity between generated text and human references. Cov measures the average % of input concepts covered by the generated text. We also introduce BERTScore Zhang et al. (2020), which measures token-by-token BERT Devlin et al. (2019) embeddings similarity. It also measures the similarity between the generated text and human references, but on a more semantic (rather than surface token) level. When reporting BERTScore, we multiply by 100. For all metrics, higher corresponds to better performance.
3 Initial Analysis
3.1 Correlation Study
We begin by conducting an analysis of the four baselines implemented and discussed in §2.2, which we refer to henceforth as BART-base-BL, BART-large-BL, T5-base-BL, and T5-large-BL. One aspect we were interested in is whether the number of input concepts affects the quality of generated text. We conduct a comprehensive correlation study of the performance of the four baselines on devO w.r.t. the number of input concepts.
As seen from Table 3, the majority of the metrics are positively correlated with concept set size across the models. ROUGE-L, CIDEr, and SPICE have small correlations that are mainly statistically insignificant, demonstrating that they are likely uncorrelated with concept set size. Coverage is strongly negatively correlated, showing that there is a higher probability of concepts missing from the generated text as concept set size increases.
There are two major takeaways from this. Firstly, increased concept set size results in greater overall performance. Secondly, models have difficulty with coverage given increased concept set size. This motivates our first set of improvements, which involves augmenting the concept sets with additional words in hopes of 1) increasing performance of the models and 2) improving their coverage, as we hope that training with more input concepts will help models learn to better cover them in the generated text. This is discussed more in §4.1.
BART-base | BART-large | T5-base | T5-large | |||||||||
Correlation | PCC | PCC | PCC | PCC | ||||||||
ROUGE-1 | 0.08 | 0.09 | 0.07 | 0.10 | 0.12 | 0.09 | 0.04 | 0.05 | 0.04 | 0.10 | 0.11 | 0.09 |
ROUGE-2 | 0.05 | 0.08 | 0.07 | 0.05 | 0.10 | 0.07 | 0.03 | 0.07 | 0.05 | 0.06 | 0.09 | 0.07 |
ROUGE-L | 0.00* | 0.01* | 0.01* | 0.00* | 0.02* | 0.01* | -0.03 | -0.01* | -0.01* | 0.02* | 0.04 | 0.03 |
BLEU-1 | 0.08 | 0.08 | 0.06 | 0.14 | 0.14 | 0.11 | 0.00* | 0.03* | 0.02* | 0.08 | 0.11 | 0.09 |
BLEU-2 | 0.06 | 0.06 | 0.04 | 0.11 | 0.11 | 0.08 | 0.03* | 0.04* | 0.03* | 0.09 | 0.10 | 0.07 |
BLEU-3 | 0.08 | 0.06 | 0.05 | 0.09 | 0.09 | 0.06 | 0.04* | 0.03* | 0.02* | 0.09 | 0.08 | 0.06 |
BLEU-4 | 0.05 | 0.05 | 0.04 | 0.05 | 0.07 | 0.05 | 0.04* | 0.02* | 0.02* | 0.08 | 0.08 | 0.06 |
METEOR | 0.05 | 0.08 | 0.06 | 0.06 | 0.09 | 0.07 | 0.02* | 0.04 | 0.03 | 0.06 | 0.08 | 0.06 |
CIDEr | -0.02* | -0.03* | -0.02* | 0.01* | 0.02* | 0.02* | -0.08 | -0.10 | -0.07 | 0.00* | 0.00* | 0.00* |
SPICE | -0.02* | -0.01* | -0.01* | 0.01* | 0.02* | 0.01* | -0.02* | -0.02* | -0.02* | 0.02* | 0.03* | 0.02* |
BERTScore | 0.04 | 0.03 | 0.02 | 0.06 | 0.06 | 0.05 | 0.04 | 0.03 | 0.02 | 0.05 | 0.04 | 0.03 |
Coverage | -0.26 | -0.31 | -0.27 | -0.07 | -0.13 | -0.11 | -0.38 | -0.42 | -0.37 | -0.26 | -0.31 | -0.28 |
3.2 Qualitative Analysis
We conduct a qualitative analysis of the baseline model outputs. We observe that several outputs are more like phrases than full coherent sentences, e.g. “body of water on a raft”. Some generated texts are also missing important words, e.g. “A listening music and dancing in a dark room” is clearly missing a noun before listening. A large portion of generated texts are quite generic and bland, e.g. “Someone sits and listens to someone talk”, while more detailed and specific statements are present in the human references. This can be seen as an instance of the noted “dull response” problem faced by generation models Du and Black (2019); Li et al. (2016), where they prefer safe, short, and frequent responses independent of the input.
Another issue is the way sentences are pieced together. Certain phrases in the outputs are either joined in the wrong order or with incorrect connectors, leading to sentences that appear to lack commonsense. For example, “body of water on a raft” is illogical, and the phrases “body of water” and “a raft” are pieced together incorrectly. Example corrections include “body of water carrying a raft” and “a raft on a body of water”. The first changes the adverb on joining them to the verb carrying, and the second pieces them together in the opposite order. A similar issue occurs with the {horse, carriage, draw} example in Table 4.
Some major takeaways are that many generations are: 1) phrases rather than full sentences and 2) poorly pieced together and lack fluency and logic compared to human references. This motivates our second set of improvements, which involves recombining extracted phrases from baseline generations into hopefully more fluent and logical sentences. This is discussed more in §4.2.
Concept Set | Baseline Generation | Human Reference |
{horse, carriage, draw} | horse drawn in a carriage | The carriage is drawn by the horse. |
{fish, catch, pole} | fish caught on a pole | The man used a fishing pole to catch fish. |
{listen, talk, sit} | Someone sits and listens to someone talk. | The man told the boy to sit down |
and listen to him talk. | ||
{bathtub, bath, dog, give} | A dog giving a bath in a bathtub. | The teenager made a big mess in the |
bathtub giving her dog a bath. |
4 SAPPHIRE Methodology
4.1 Concept Set Augmentation
The first set of improvements is concept set augmentation, which involves augmenting the input concept sets. We try augmentation using up to 1 to 5 additional words, and train-time augmentation both with and without test-time augmentation. We observed that test-time augmentation resulted in inconsistent results that were not as effective, and stick with train-time only augmentation. During training, rather than feeding in the original concept sets as inputs, we instead feed in these augmented concept sets which consist of more words. The expected outputs are the same human references. During test-time, we simply feed in the original concept sets (without augmentation) as inputs.
4.1.1 Keyword-based Augmentation
The first type of augmentation we try is keyword-based, or Kw-aug. We augment the trainCG concept sets with keywords extracted from the human references using KeyBERT333https://github.com/MaartenGr/KeyBERT Grootendorst (2020). We calculate the average semantic similarity (using cosine similarity of BERT embeddings) of the candidate keywords to the original concept set. At each stage of augmentation, we add the remaining candidate with the highest similarity.444We also tried using the least semantically similar keywords, but results were noticeably worse. Some augmentation examples can be found in Table 5. We train our BART and T5 models using these augmented sets, and call the resulting models BART-base-KW, BART-large-KW, T5-base-KW, and T5-large-KW.
Method | Original Concept Set | Added Words |
Kw-aug | {match, stadium, watch} | {soccer, league, fans} |
Kw-aug | {family, time, spend} | {holidays} |
Kw-aug | {head, skier, slope} | {cabin} |
Att-aug | {boat, lake, drive} | {fisherman} |
Att-aug | {family, time, spend} | {at, holidays} |
Att-aug | {player, match, look} | {tennis, on, during} |
4.1.2 Attention-based Augmentation
We also try attention-based augmentation, or Att-aug. We augment the trainCG concept sets with the words that have been most attended upon in aggregate by the other words in the human references. We pass the reference texts through BERT and return the attention weights at the last layer. At each stage of augmentation, we add the remaining candidate word with the highest attention. Adding the least attended words would not be effective as many are words with little meaning (e.g. simple articles such as “a” and “the”). Some augmentation examples can be found in Table 5. We train our BART and T5 models using these augmented sets, and call the resulting models BART-base-Att, BART-large-Att, T5-base-Att, and T5-large-Att.
4.2 Phrase Recombination
The second set of improvements is phrase recombination, which involves breaking down sentences into phrases and reconstructing them into new sentences which are hopefully more logical and coherent. For training, we use YAKE Campos et al. (2018) to break down the trainCG human references into phrases of up to 2, 3, and 5 n-grams long, and ensure extracted phrases have as little overlap as possible. During validation and testing, since we assume no access to ground-truth human references, we instead use YAKE to extract keyphrases from our baseline model generations.
We ignore extracted 1-grams as this approach focuses on phrases. We find words from the original concept set which are not covered by our extracted keyphrases and include them to ensure that coverage is maintained. Essentially, we form a new concept set which can also consist of phrases. Some examples can be found in Table 6.
Original Text | Extracted Keyphrases | New Input Concept Set |
A dog wags his tail at the boy. | dog wags his tail | {dog wags his tail} |
hanging a painting on a wall at home | hanging a painting | {hanging a painting, wall} |
a herd of many sheep crowded together in a stable | herd of many sheep crowded | {herd of many sheep crowded, |
waiting to be dipped for ticks and other pests | dip, waiting} | |
a soldier takes a knee while providing security | knee while providing security, | {knee while providing security, |
during a patrol outside of the village. | patrol outside of the village | patrol outside of the village, take} |
4.2.1 Phrase-to-text (P2T)
To piece the phrases back together, we try phrase-to-text (P2T) generation by training BART and T5 to generate full sentences given our new input sets, and call these models BART-base-P2T, BART-large-P2T, T5-base-P2T, and T5-large-P2T. During training, we choose a single random permutation of each training input set (consisting of extracted keyphrases from the human references), with the elements separated by s, and the human references as the outputs. This is in order for the models to learn to be order-agnostic, which is important as one desired property of phrase recombination is the ability to combine phrases in different orders, as motivated by the qualitative analysis in §3.2. During inference or test-time, we feed in a single random permutation of each test-time input set, consisting of extracted keyphrases from the corresponding baseline model’s outputs.
4.2.2 Mask Infilling (MI)
This method interpolates text between test-time input set elements with no training required. For example, given a test-time input set {,}, we feed in “mask mask mask” and “mask mask mask” to an MI model to fill the mask tokens with text. We use BART-base and BART-large for MI, and call the approaches BART-base-MI and BART-large-MI, respectively. We use BART-base-MI on input sets consisting of extracted keyphrases from BART-base-BL and T5-base-BL, and BART-large-MI on input sets consisting of extracted keyphrases from BART-large-BL and T5-large-BL. We also try MI on the original concept sets (with no phrases).
One difficulty is determining the right input set permutation. Many contain 5 elements (meaning 5!=120 permutations), making exhaustive MI infeasible. Order of elements for infilling can result in vastly different outputs (see §6.3), as certain orders are more natural. Humans perform their own intuitive reordering of given inputs when writing, and the baselines and other approaches (e.g. Kw-aug, P2T) learn to mainly be order agnostic.
We use perplexity (PPL) from GPT-2 Radford et al. (2019) to pick the “best” permutations for MI. We feed up to 120 permutations of each input set (with elements separated by spaces) to GPT-2 to extract their PPL, and keep the 10 with lowest PPL per example. This is not a perfect approach, but is likely better than random sampling. For each example, we perform MI on these ten permutations, and select the output with lowest GPT-2 PPL.
We found BART-large-MI outputs contain URLs, news agency names in brackets, etc. Hence, we post-process before output selection and evaluation. BART-base-MI does not do this. One possible explanation is that BART-large may have been pretrained on more social media and news data.
5 Experiments
5.1 Model Training and Selection
For training Kw-aug, Att-aug, and P2T models, we follow baseline hyperparameters, barring learning rate (LR) which is tuned per-method. We train two seeds per model. See Appendix A for more.
For each model, we choose the epoch corresponding to highest ROUGE-2 on the dev split, and use beam search for decoding. The dev and test splits are different. For Kw-aug and Att-aug models, the splits are simply devCG and testCG (or testO), as we do not perform test-time augmentation. For P2T, the splits are devCG and testCG (or testO) but with the input sets replaced with new ones that include keyphrases extracted from the corresponding baseline model’s outputs.
The number of words to augment for Kw-aug and Att-aug (from 1 to 5) and maximum n-gram length of extracted keyphrases for P2T (2, 3, or 5) are hyperparameters. While we train separate versions of each model corresponding to different values of these, the final chosen model per method and model combination (such as BART-base-KW) is the one corresponding to the hyperparameter value that performs best on the dev split when averaged over both seeds. For MI, which involves no training, we select the variation (MI on the original concept set or new input sets with keyphrases up to 2, 3, or 5 n-grams) per model that performs best on the dev split, and only perform infilling using extracted keyphrases from the first seed baseline generations. These are the selected models we report the testCG and testO results of in §6.
5.2 Human Evaluation
We ask annotators to evaluate 48 testCG examples from the human references, baseline outputs, and various method (excluding MI) outputs for BART-large and T5-base. We choose these two as they cover both model types and sizes, and exclude MI as it performs noticeably worse on the automatic evaluation (see §6.1). See Appendix §C for more.
The annotators evaluate fluency and commonsense of the texts on 1-5 scales. Fluency, also known as naturalness, is a measure of how human-like a text is. Commonsense is the plausibility of the events described. We do not evaluate coverage as automatic metrics suffice; coverage is more objective compared to fluency and commonsense.
BART-base | |||||
Metrics\Methods | Baseline | Kw-aug | Att-aug | P2T | BART-base-MI |
ROUGE-1 | 43.96 | 45.01 | 44.99 | 44.87 | 44.83 |
ROUGE-2 | 17.31 | 18.33 | 18.18 | 18.04 | 17.44 |
ROUGE-L | 36.65 | 37.28 | 37.76 | 37.28 | 34.47 |
BLEU-1 | 73.20 | 73.00 | 73.00 | 73.15 | 69.90 |
BLEU-2 | 54.50 | 55.35 | 55.70 | 55.65 | 49.00 |
BLEU-3 | 40.40 | 41.35 | 41.40 | 41.85 | 34.70 |
BLEU-4 | 30.10 | 31.10 | 30.95 | 31.75 | 24.70 |
METEOR | 30.35 | 30.50 | 30.70 | 31.05 | 29.70 |
CIDEr | 15.56 | 16.18 | 15.68 | 16.14 | 14.43 |
SPICE | 30.05 | 30.45 | 30.65 | 30.95 | 28.40 |
BERTScore | 59.19 | 59.32 | 59.72 | 59.54 | 53.73 |
Coverage | 90.43 | 91.44 | 91.23 | 91.47 | 96.23 |
BART-large | |||||
Metrics\Methods | Baseline | Kw-aug | Att-aug | P2T | BART-large-MI |
ROUGE-1 | 45.67 | 46.71 | 46.78 | 46.26 | 41.69 |
ROUGE-2 | 18.77 | 19.64 | 19.92 | 19.37 | 15.40 |
ROUGE-L | 37.83 | 38.38 | 38.53 | 38.22 | 33.32 |
BLEU-1 | 74.45 | 76.20 | 76.55 | 77.10 | 63.90 |
BLEU-2 | 56.25 | 58.60 | 59.60 | 58.95 | 42.40 |
BLEU-3 | 42.15 | 44.00 | 45.20 | 44.70 | 29.20 |
BLEU-4 | 32.10 | 33.40 | 34.50 | 34.25 | 20.50 |
METEOR | 31.70 | 32.60 | 32.65 | 33.00 | 28.30 |
CIDEr | 16.42 | 17.37 | 17.49 | 17.50 | 12.32 |
SPICE | 31.85 | 33.15 | 33.30 | 33.60 | 26.10 |
BERTScore | 59.95 | 60.83 | 60.87 | 61.30 | 48.56 |
Coverage | 94.49 | 96.74 | 96.02 | 97.02 | 95.33 |
T5-base | |||||
Metrics\Methods | Baseline | Kw-aug | Att-aug | P2T | BART-base-MI |
ROUGE-1 | 44.63 | 46.42 | 46.750.11 | 45.73 | 44.92 |
ROUGE-2 | 18.40 | 19.36 | 19.200.17 | 18.51 | 17.98 |
ROUGE-L | 37.60 | 38.68 | 38.510.21 | 38.07 | 34.88 |
BLEU-1 | 73.60 | 76.25 | 76.000.28 | 75.65 | 70.20 |
BLEU-2 | 57.00 | 59.55 | 58.750.35 | 58.15 | 50.50 |
BLEU-3 | 42.75 | 45.10 | 44.000.28 | 43.45 | 36.20 |
BLEU-4 | 32.70 | 34.45 | 33.300.28 | 33.10 | 26.10 |
METEOR | 31.05 | 31.85 | 31.900.14 | 32.05 | 30.20 |
CIDEr | 16.26 | 17.37 | 17.040.21 | 16.84 | 14.83 |
SPICE | 31.95 | 32.75 | 32.850.21 | 33.20 | 29.70 |
BERTScore | 61.40 | 61.88 | 61.280.10 | 61.46 | 55.04 |
Coverage | 90.96 | 94.92 | 96.000.03 | 94.78 | 96.03 |
T5-large | |||||
Metrics\Methods | Baseline | Kw-aug | Att-aug | P2T | BART-large-MI |
ROUGE-1 | 46.26 | 47.47 | 47.40 | 46.72 | 42.78 |
ROUGE-2 | 19.62 | 20.02 | 20.19 | 19.76 | 16.61 |
ROUGE-L | 39.21 | 39.84 | 39.97 | 39.19 | 34.52 |
BLEU-1 | 77.45 | 78.70 | 78.95 | 77.90 | 66.80 |
BLEU-2 | 60.75 | 62.10 | 62.35 | 61.00 | 45.90 |
BLEU-3 | 46.60 | 47.65 | 47.95 | 46.75 | 32.70 |
BLEU-4 | 36.30 | 36.80 | 37.35 | 36.10 | 23.90 |
METEOR | 33.30 | 33.55 | 33.70 | 33.35 | 29.10 |
CIDEr | 17.90 | 18.40 | 18.43 | 17.89 | 13.34 |
SPICE | 34.25 | 34.50 | 33.70 | 34.00 | 28.00 |
BERTScore | 62.65 | 62.91 | 62.78 | 62.46 | 50.57 |
Coverage | 94.23 | 95.92 | 96.08 | 95.44 | 96.03 |
6 Results and Analysis
BART-base | BART-large | T5-base | T5-large | ||
p-values | P2T | Att-aug | P2T | Kw-aug | Att-aug |
ROUGE-1 | 1.58E-05 | 1.58E-05 | 7.58E-04 | 1.58E-05 | 1.58E-05 |
ROUGE-2 | 6.32E-05 | 1.58E-05 | 2.18E-03 | 1.58E-05 | 2.20E-03 |
ROUGE-L | 6.32E-05 | 8.53E-04 | 2.78E-02 | 1.58E-05 | 1.58E-05 |
BLEU-1 | 3.63E-01 | 1.39E-04 | 6.94E-05 | 6.94E-05 | 1.11E-03 |
BLEU-2 | 1.11E-03 | 6.94E-05 | 6.94E-05 | 6.94E-05 | 5.69E-03 |
BLEU-3 | 3.26E-02 | 1.04E-03 | 9.03E-04 | 4.17E-04 | 3.40E-02 |
BLEU-4 | 5.68E-02 | 1.57E-01 | 8.40E-03 | 1.83E-02 | 2.66E-01 |
METEOR | 1.57E-02 | 9.03E-04 | 6.94E-05 | 2.08E-04 | 7.27E-01 |
CIDEr | 6.25E-04 | 2.08E-04 | 6.94E-05 | 6.94E-05 | 5.07E-03 |
SPICE | 1.53E-03 | 6.25E-04 | 6.94E-05 | 1.43E-02 | 9.16E-01 |
BERTScore | 3.33E-03 | 1.58E-05 | 1.58E-05 | 1.58E-05 | 1.42E-01 |
Coverage | 3.16E-05 | 1.58E-05 | 1.58E-05 | 1.58E-05 | 1.58E-05 |
Models\Metrics | ROUGE-2/L | BLEU-3/4 | METEOR | CIDEr | SPICE | Coverage | ||
T5-base (reported baseline) | 14.63 | 34.56 | 28.76 | 18.54 | 23.94 | 9.40 | 19.87 | 76.67 |
BART-large (reported baseline) | 22.02 | 41.78 | 39.52 | 29.01 | 31.83 | 13.98 | 28.00 | 97.35 |
T5-large (reported baseline) | 21.74 | 42.75 | 43.01 | 31.96 | 31.12 | 15.13 | 28.86 | 95.29 |
EKI-BART Fan et al. (2020) | - | - | - | 35.945 | - | 16.999 | 29.583 | - |
KG-BART Liu et al. (2021) | - | - | - | 33.867 | - | 16.927 | 29.634 | - |
RE-T5 Wang et al. (2021) | - | - | - | 40.863 | - | 17.663 | 31.079 | - |
BART-base-P2T | 20.83 | 42.91 | 40.74 | 29.918 | 30.61 | 14.670 | 26.960 | 92.84 |
T5-base-P2T | 22.38 | 44.59 | 44.97 | 33.577 | 31.95 | 16.152 | 29.104 | 95.55 |
BART-large-KW | 22.25 | 43.38 | 43.87 | 32.956 | 32.26 | 16.065 | 28.335 | 96.16 |
BART-large-Att | 22.22 | 43.80 | 44.61 | 33.405 | 32.03 | 16.036 | 28.452 | 96.43 |
BART-large-P2T | 22.65 | 43.84 | 44.78 | 33.961 | 32.18 | 16.174 | 28.462 | 96.20 |
T5-large-KW | 23.79 | 45.73 | 48.06 | 37.023 | 32.85 | 16.987 | 29.659 | 95.32 |
T5-large-Att | 23.94 | 45.87 | 47.99 | 36.947 | 32.79 | 16.943 | 29.607 | 95.43 |
T5-large-P2T | 23.89 | 45.77 | 48.08 | 37.119 | 32.94 | 16.901 | 29.751 | 94.82 |
Model | Method | Fluency | Commonsense |
BART-large | Baseline | 3.92 | 4.06 |
Kw-aug | 4.13 | 3.92 | |
Att-aug | 4.10 | 4.06 | |
P2T | 4.17 | 4.13 | |
T5-base | Baseline | 4.02 | 3.83 |
Kw-aug | 4.04 | 4.04 | |
Att-aug | 4.13 | 3.98 | |
P2T | 4.02 | 4.08 | |
Human | 4.14 | 4.32 |
Automatic evaluation results on testCG can be found in Tables 7, 8, 9, 10, and results on testO in Table 12. Human evaluation results on testCG can be found in Table 13. Single keyword augmentation performs best for Kw-aug across models. Two word augmentation performs best for Att-aug, except T5-base where three word augmentation performs best. Keyphrases up to 2-grams long perform best for P2T, except T5-large where 3-grams perform best. All models perform best with keyphrases up to 5-grams long for MI. These are the results reported here, and graphs displaying other hyperparameter results on testCG are in Appendix D. Table 14 contains qualitative examples, and more can be found in Appendix §E.
6.1 Automatic Evaluation
We see from Tables 7 to 10 that SAPPHIRE methods outperform the baselines on most/all metrics across the models on testCG. The only exception is MI, which performs worse other than coverage.
For BART-base, Kw-aug, Att-aug, and P2T all outperform the baseline across the metrics. For BART-large, Att-aug and P2T outperform the baseline heaviest, with noticeable increases to all metrics. For T5-base, all methods outperform the baseline, with Kw-aug performing best. Att-aug performs best for T5-large, and SAPPHIRE appears relatively less effective for T5-large. T5-large is the strongest baseline, and hence further improving its performance is possibly more difficult.
MI performs worse across most metrics except coverage, likely as MI almost always keeps inputs intact in their exact form. This is however possibly one reason for its low performance; it is less flexible. Further, as discussed in §4.2.2, MI is highly dependent on the input order. See §6.3 for more.
Table 11 contains statistical significance p-values from Pitman’s permutation tests Pitman (1937) for what we adjudged to be the best performing method(s) per model compared to corresponding baselines on testCG. Most metrics across the methods are significant compared to the baselines.
From Table 12, we see that SAPPHIRE models outperform the corresponding baselines reported in Lin et al. (2020) on testO. T5-large-KW and P2T outperform EKI-BART Fan et al. (2020) and KG-BART Liu et al. (2021) on both SPICE and BLEU-4, which are two SOTA published CommonGen models that use external knowledge from corpora and KGs. As SPICE is used to rank the CommonGen leaderboard555https://inklab.usc.edu/CommonGen/leaderboard.html, T5-large-KW and P2T would place highly. SAPPHIRE models do lag behind the SOTA published RE-T5 Wang et al. (2021), showing potential for further improvement. Further, the BART-large SAPPHIRE models perform worse than EKI-BART and KG-BART, but not by a substantial margin. We emphasize again that SAPPHIRE simply uses the data itself and the baseline generations, rather than external knowledge. Hence, SAPPHIRE’s performance gains over the baselines and certain SAPPHIRE models matching or outperforming SOTA models that leverage external information is quite impressive.
6.2 Human Evaluation
Table 13 shows human evaluation results on testCG for human references and methods (excluding MI) using BART-large and T5-base. SAPPHIRE generally outperforms the baselines. BART-large-P2T performs noticeably higher on both fluency and commonsense. For T5-base, all three methods outperform the baseline across both metrics. Compared to humans, our best methods have comparable fluency, but still lag noticeably on commonsense, demonstrating that human-level generative commonsense reasoning is indeed challenging.
6.3 Qualitative Analysis
We see from Table 14 that many baseline outputs contain issues found in §3.2, e.g. incomplete or illogical sentences. Human references are fluent, logical, and sometimes more creative (e.g. example 5), which all methods still lack in comparison.
For example 1, the baseline generation “hands sitting on a chair” misses the concept “toy”, whereas our methods do not. Kw-aug and Att-aug output complete and logical sentences. For example 2, the baseline generation of “a camel rides a camel” is illogical. Our methods output more logical and specific sentences. For example 3, our methods generate more complete and coherent sentences than the baseline, which lacks a subject (does not mention who is “walking”). For example 4, the baseline generation “bus sits on the tracks” is illogical as buses park on roads. Our methods do not suffer from this and output more reasonable text. For example 5, the baseline generation “A lady sits in a sunglass.” is completely illogical. Kw-aug, Att-aug, and P2T all output logical text. For example 6, the baseline output “Someone stands in front of someone holding a hand” is generic and bland. Kw-aug, Att-aug, and P2T all output more specific and detailed text rather than simply referring to “someone”. Overall, SAPPHIRE generates text that is more complete, fluent, logical, and with greater coverage, addressing many baseline issues (§3.2).
However, SAPPHIRE methods are imperfect. P2T relies heavily on the original generation. For example 1, the baseline output “hands sitting on a chair” is extracted as a keyphrase, and used in the P2T output “hands sitting on a chair with toys”. While coverage improves, the text is still illogical. For example 2, P2T still misses the “walk” concept. While the Att-aug output of “A man is riding camel as he walks through the desert.” is more logical than the baseline’s, it is still not entirely logical as the man cannot ride the camel and walk at the same time. MI outputs logical and fluent text for examples 2 and 3. For the other examples, the generated texts are illogical, not fluent, or incomplete.
This is likely due to input permutation having a strong effect on output quality. For example, “wave” before “falls off a surf board” leads to an illogical output “A wave falls off a surf board.”, where the reverse order results in a more logical output “A man falls off a surf board and hits a wave.” As discussed in §4.2.2, our method of selecting best permutations is likely imperfect. Further, BART-MI usually does not inflect inputs and retains them in exact form, unlike the baselines and other methods which learn to inflect words (e.g. singular to plural). We believe BART-MI has potential if these weaknesses can be addressed.
Method | Text |
Concept Set | {sit, chair, toy, hand} (example 1) |
BART-base-BL | hands sitting on a chair |
BART-base-KW | A boy sits on a chair with a toy in his hand. |
BART-base-Att | A child sits on a chair with a toy in his hand. |
BART-base-P2T | hands sitting on a chair with toys |
BART-base-MI | Children’s hands sit on a chair with a toy. |
Human | A baby sits on a chair with a toy in one of its hands. |
Concept Set | {camel, desert, ride, walk} (example 2) |
BART-base-BL | a camel rides a camel in the desert |
BART-base-KW | A camel rides down a walkway in the desert. |
BART-base-Att | A man is riding camel as he walks through the desert. |
BART-base-P2T | A camel rides down a trail in the desert. |
BART-base-MI | In the desert, a man rides a camel for a walk. |
Human | A loud group of people walk around the desert and ride camels. |
Concept Set | {jacket, wear, snow, walk} (example 3) |
BART-large-BL | walking in the snow wearing a furry jacket |
BART-large-KW | A man wearing a jacket is walking in the snow. |
BART-large-Att | A man in a blue jacket is walking in the snow. |
BART-large-P2T | A man is wearing a furry jacket as he walks in the snow. |
BART-large-MI | A walk in the snow wearing a furry jacket |
Human | A man wears a jacket and walks in the snow. |
Concept Set | {bench, bus, wait, sit} (example 4) |
BART-large-BL | A bus sits on the tracks with people waiting on benches. |
BART-large-KW | A bus sits next to a bench waiting for passengers. |
BART-large-Att | A woman sits on a bench waiting for a bus. |
BART-large-P2T | A bus sits at a stop waiting for passengers to get off the bench. |
BART-large-MI | There are people waiting on benches outside bus stops |
to sit down. pic.twitter. | |
Human | The man sat on the bench waiting for the bus. |
Concept Set | {sunglass, wear, lady, sit} (example 5) |
T5-base-BL | A lady sits in a sunglass. |
T5-base-KW | A lady sits next to a man wearing sunglasses. |
T5-base-Att | A lady sits wearing sunglasses. |
T5-base-P2T | A lady sits next to a man wearing sunglasses. |
BART-base-MI | A young lady sits in a sunglass to wear. |
Human | The lady wants to wear sunglasses, sit, relax, |
and enjoy her afternoon. | |
Concept Set | {hold, hand, stand, front} (example 6) |
T5-large-BL | Someone stands in front of someone holding a hand. |
T5-large-KW | Two men stand in front of each other holding hands. |
T5-large-Att | A man stands in front of a woman holding a hand. |
T5-large-P2T | A man standing in front of a man holding a hand. |
BART-large-MI | Mr. Trump holding a hand to stand in front of |
Human | A man stands and holds his hands out in front of him. |
7 Related Work
Constrained Text Generation:
There has been more work on constrained text generation. Miao et al. (2019) use Metropolis-Hastings sampling to determine token-level edits at each step of generation. Feng et al. (2019) introduce Semantic Text Exchange to adjust the semantics of a text given a replacement entity. Gangal et al. (2021a) propose narrative reordering (NAREOR) to rewrite stories in different narrative orders while preserving plot.
Data-to-text NLG:
A wide range of data-to-text NLG benchmarks have been proposed, e.g. for generating weather reports Liang et al. (2009), game commentary Jhamtani et al. (2018), and recipes Kiddon et al. (2016). E2E-NLG Dušek et al. (2018) and WebNLG Gardent et al. (2017) are two benchmarks that involve generating text from meaning representation (MR) and triple sequences. Montella et al. (2020) use target Wiki sentences with parsed OpenIE triples as weak supervision for WebNLG. Tandon et al. (2018) permute input MRs to augment examples for E2E-NLG.
Commonsense Reasoning and Incorporation:
Talmor et al. (2020) show that not all pretrained LMs can reason through commonsense tasks. Other works investigate commonsense injection into models; one popular way is through knowledge graphs (KGs). One large commonsense KG is COMET, which trains on KG edges to learn connections between words and phrases. COSMIC Ghosal et al. (2020) uses COMET to inject commonsense. EKI-BART Fan et al. (2020) and KG-BART Liu et al. (2021) show that external knowledge (from corpora and KGs) can improve performance on CommonGen. Distinctly, SAPPHIRE obviates reliance on external knowledge.
8 Conclusion and Future Work
In conclusion, we motivated and proposed several improvements for concept-to-text generation which we call SAPPHIRE: Set Augmentation and Post-hoc PHrase Infilling and REcombination. We demonstrated their effectiveness on CommonGen through experiments on BART and T5. Extensive evaluation showed that SAPPHIRE improves model performance, addresses many issues of the baselines, and has potential for further exploration.
Potential future work includes improving mask infilling performance, and trying combinations of SAPPHIRE techniques as they could be complementary. Better exploiting regularities of CommonGen-like tasks, e.g. invariance to input order, presents another avenue. SAPPHIRE methods can also be investigated for other data-to-text NLG tasks, e.g. WebNLG, and explored for applications such as improving the commonsense reasoning of personalized dialog agents Li et al. (2020), data augmentation for NLG Feng et al. (2020, 2021), and constructing pseudo-references for long-context NLG Gangal et al. (2021b).
Acknowledgments
We thank our anonymous reviewers, Graham Neubig, Ritam Dutt, Divyansh Kaushik, and Zhengbao Jiang for their comments and suggestions.
References
- Anderson et al. (2016) Peter Anderson, Basura Fernando, Mark Johnson, and Stephen Gould. 2016. Spice: Semantic propositional image caption evaluation. In European conference on computer vision, pages 382–398. Springer.
- Bosselut et al. (2019) Antoine Bosselut, Hannah Rashkin, Maarten Sap, Chaitanya Malaviya, Asli Celikyilmaz, and Yejin Choi. 2019. COMET: Commonsense transformers for automatic knowledge graph construction. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 4762–4779, Florence, Italy. Association for Computational Linguistics.
- Campos et al. (2018) Ricardo Campos, Vítor Mangaravite, Arian Pasquali, A. Jorge, C. Nunes, and A. Jatowt. 2018. Yake! collection-independent automatic keyword extractor. In ECIR.
- Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics.
- Du and Black (2019) Wenchao Du and Alan W Black. 2019. Boosting dialog response generation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 38–43, Florence, Italy. Association for Computational Linguistics.
- Dušek et al. (2018) Ondřej Dušek, Jekaterina Novikova, and Verena Rieser. 2018. Findings of the E2E NLG challenge. In Proceedings of the 11th International Conference on Natural Language Generation, pages 322–328, Tilburg University, The Netherlands. Association for Computational Linguistics.
- Fan et al. (2020) Zhihao Fan, Yeyun Gong, Zhongyu Wei, Siyuan Wang, Yameng Huang, Jian Jiao, Xuanjing Huang, Nan Duan, and Ruofei Zhang. 2020. An enhanced knowledge injection model for commonsense generation. In Proceedings of the 28th International Conference on Computational Linguistics, pages 2014–2025, Barcelona, Spain (Online). International Committee on Computational Linguistics.
- Feng et al. (2020) Steven Y. Feng, Varun Gangal, Dongyeop Kang, Teruko Mitamura, and Eduard Hovy. 2020. GenAug: Data augmentation for finetuning text generators. In Proceedings of Deep Learning Inside Out (DeeLIO): The First Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, pages 29–42, Online. Association for Computational Linguistics.
- Feng et al. (2021) Steven Y. Feng, Varun Gangal, Jason Wei, Sarath Chandar, Soroush Vosoughi, Teruko Mitamura, and Eduard Hovy. 2021. A survey of data augmentation approaches for NLP. Association for Computational Linguistics (ACL) 2021 Findings.
- Feng et al. (2019) Steven Y. Feng, Aaron W. Li, and Jesse Hoey. 2019. Keep calm and switch on! preserving sentiment and fluency in semantic text exchange. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 2701–2711.
- Gangal et al. (2021a) Varun Gangal, Steven Y. Feng, Eduard Hovy, and Teruko Mitamura. 2021a. NAREOR: The narrative reordering problem. arXiv preprint arXiv:2104.06669.
- Gangal et al. (2021b) Varun Gangal, Harsh Jhamtani, Eduard Hovy, and Taylor Berg-Kirkpatrick. 2021b. Improving automated evaluation of open domain dialog via diverse reference augmentation. arXiv preprint arXiv:2106.02833.
- Gardent et al. (2017) Claire Gardent, Anastasia Shimorina, Shashi Narayan, and Laura Perez-Beltrachini. 2017. The webnlg challenge: Generating text from rdf data. In Proceedings of the 10th International Conference on Natural Language Generation, pages 124–133.
- Ghosal et al. (2020) Deepanway Ghosal, Navonil Majumder, Alexander Gelbukh, Rada Mihalcea, and Soujanya Poria. 2020. COSMIC: COmmonSense knowledge for eMotion identification in conversations. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 2470–2481, Online. Association for Computational Linguistics.
- Grootendorst (2020) Maarten Grootendorst. 2020. Keybert: Minimal keyword extraction with bert.
- Jhamtani et al. (2018) Harsh Jhamtani, Varun Gangal, Eduard Hovy, Graham Neubig, and Taylor Berg-Kirkpatrick. 2018. Learning to generate move-by-move commentary for chess games from large-scale social forum data. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1661–1671.
- Kiddon et al. (2016) Chloé Kiddon, Luke Zettlemoyer, and Yejin Choi. 2016. Globally coherent text generation with neural checklist models. In Proceedings of the 2016 conference on empirical methods in natural language processing, pages 329–339.
- Lewis et al. (2020) Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Veselin Stoyanov, and Luke Zettlemoyer. 2020. BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 7871–7880, Online. Association for Computational Linguistics.
- Li et al. (2020) Aaron W. Li, Veronica Jiang, Steven Y. Feng, Julia Sprague, Wei Zhou, and Jesse Hoey. 2020. ALOHA: Artificial learning of human attributes for dialogue agents. Proceedings of the AAAI Conference on Artificial Intelligence, 34(05):8155–8163.
- Li et al. (2016) Jiwei Li, Michel Galley, Chris Brockett, Jianfeng Gao, and Bill Dolan. 2016. A diversity-promoting objective function for neural conversation models. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 110–119, San Diego, California. Association for Computational Linguistics.
- Liang et al. (2009) Percy Liang, Michael I Jordan, and Dan Klein. 2009. Learning semantic correspondences with less supervision. In Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP, pages 91–99.
- Lin et al. (2020) Bill Yuchen Lin, Wangchunshu Zhou, Ming Shen, Pei Zhou, Chandra Bhagavatula, Yejin Choi, and Xiang Ren. 2020. CommonGen: A constrained text generation challenge for generative commonsense reasoning. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 1823–1840, Online. Association for Computational Linguistics.
- Liu et al. (2021) Ye Liu, Yao Wan, Lifang He, Hao Peng, and Philip S. Yu. 2021. KG-BART: Knowledge graph-augmented bart for generative commonsense reasoning. Proceedings of the AAAI Conference on Artificial Intelligence, 35(7):6418–6425.
- Miao et al. (2019) Ning Miao, Hao Zhou, Lili Mou, Rui Yan, and Lei Li. 2019. CGMH: Constrained sentence generation by metropolis-hastings sampling. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 6834–6842.
- Montella et al. (2020) Sebastien Montella, Betty Fabre, Tanguy Urvoy, Johannes Heinecke, and Lina Rojas-Barahona. 2020. Denoising pre-training and data augmentation strategies for enhanced RDF verbalization with transformers. In Proceedings of the 3rd International Workshop on Natural Language Generation from the Semantic Web (WebNLG+), pages 89–99, Dublin, Ireland (Virtual). Association for Computational Linguistics.
- Papineni et al. (2002) Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. BLEU: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, pages 311–318.
- Pitman (1937) Edwin JG Pitman. 1937. Significance tests which may be applied to samples from any populations. Supplement to the Journal of the Royal Statistical Society, 4(1):119–130.
- Radford et al. (2019) Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language models are unsupervised multitask learners. OpenAI Blog, 1(8):9.
- Raffel et al. (2020) Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2020. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research, 21(140):1–67.
- Talmor et al. (2020) Alon Talmor, Yanai Elazar, Yoav Goldberg, and Jonathan Berant. 2020. oLMpics-on what language model pre-training captures. Transactions of the Association for Computational Linguistics, 8:743–758.
- Tandon et al. (2018) Shubhangi Tandon, TS Sharath, Shereen Oraby, Lena Reed, Stephanie Lukin, and Marilyn Walker. 2018. TNT-NLG, System 2: Data repetition and meaning representation manipulation to improve neural generation. E2E NLG Challenge System Descriptions.
- Vedantam et al. (2015) Ramakrishna Vedantam, C Lawrence Zitnick, and Devi Parikh. 2015. CIDEr: Consensus-based image description evaluation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4566–4575.
- Wang et al. (2021) Han Wang, Yang Liu, Chenguang Zhu, Linjun Shou, Ming Gong, Yichong Xu, and Michael Zeng. 2021. Retrieval enhanced model for commonsense generation. Association for Computational Linguistics (ACL) 2021 Findings.
- Zhang et al. (2020) Tianyi Zhang, Varsha Kishore, Felix Wu, Kilian Q Weinberger, and Yoav Artzi. 2020. BERTScore: Evaluating text generation with BERT. In International Conference on Learning Representations 2020.
Appendices
Appendix A Model Training and Generation Details
T5-large consists of 770M params, T5-base 220M params, BART-large 406M params, and BART-base 139M params. We train two seeded versions of each baseline model and SAPPHIRE model. For all models, we use beam search with a beam size of 5, decoder early stopping, a decoder length penalty of 0.6, encoder and decoder maximum lengths of 32, and a decoder minimum length of 1. For model training, we use a batch size of 128 for T5-base and BART-base, 32 for BART-large, and 16 for T5-large. For T5-base, T5-large, and BART-base, we use 400 warmup steps, and 500 for BART-large. We train all models up to a reasonable number of epochs (e.g. 10 or 20) and perform early stopping using our best judgment (e.g. if metrics have continually decreased for multiple epochs). The learning rates for SAPPHIRE models were determined by trying a range of values (e.g. from 1e-6 to 1e-4), and finding ones which led to good convergence behavior (e.g. validation metrics increase at a decently steady rate and reach max. after a reasonable number of epochs). For the best-performing models, learning rates are as follows (each set consists of {baseline,Kw-aug,Att-aug,P2T}): BART-base = {3e-05,2e-05,3e-05,1e-05}, BART-large = {3e-05,2e-05,2e-05,5e-06}, T5-base = {5e-05,5e-05,5e-05,1e-05}, T5-large = {2e-05,2e-05,2e-05,5e-06}.
Training was done using single RTX 2080 Ti and Titan Xp GPUs, and Google Colab instances which alternately used a single V100, P100, or Tesla T4 GPU. The vast majority of the training was done on a single V100 per model. T5-base models trained in approx. 1 hour, BART-base models in approx. 45 minutes, T5-large models in approx. 4 hours, and BART-large models in approx. 1.5-2 hours.
Appendix B Full Re-implementation versus Reported Model Numbers
See Table 16 for full comparison of our re-implemented CommonGen models compared to the original reported numbers in Lin et al. (2020).
Method | Text |
Concept Set | {food, eat, hand, bird} |
BART-base-BL | hands of a bird eating food |
BART-base-KW | a bird eats food from a hand |
BART-base-Att | hand of a bird eating food |
BART-base-P2T | A bird is eating food with its hand. |
BART-base-MI | The food is in the hands of a bird eating it. |
Human | A small bird eats food from someone’s hand. |
Concept Set | {front, dance, routine, perform} |
BART-base-BL | A woman performs a routine in front of a dance. |
BART-base-KW | A man performs a routine in front of a group of people. |
BART-base-Att | A man is performing a routine in front of a group of people. |
BART-base-P2T | A woman performs a routine in front of a group of people. |
BART-base-MI | In this dance, a man performs a routine in front of a mirror. |
Human | The girl performed her dance routine in front of the audience. |
Concept Set | {chase, ball, owner, dog, throw} |
BART-base-BL | A dog is throwing a ball into a chase. |
BART-base-KW | A dog is about to throw a ball to its owner. |
BART-base-Att | A dog is trying to throw a ball at its owner. |
BART-base-P2T | A dog is chasing the owner of a ball. |
BART-base-MI | The dog was trained to throw balls and the dog would chase after the owner. |
Human | The owner threw the ball for the dog to chase after. |
Concept Set | {music, dance, room, listen} |
BART-large-BL | A listening music and dancing in a dark room |
BART-large-KW | A group of people dance and listen to music in a room. |
BART-large-Att | A group of people are dancing and listening to music in a room. |
BART-large-P2T | Two people are dancing and listening to music in a dark room. |
BART-large-MI | Music and dancing in the dance floor. |
Human | A boy danced around the room while listening to music. |
Concept Set | {cheer, team, crowd, goal} |
T5-base-BL | the crowd cheered after the goal. |
T5-base-KW | the crowd cheered after the goal by football team |
T5-base-Att | the crowd cheered after the goal by the team. |
T5-base-P2T | the crowd cheered as the team scored their first goal. |
BART-base-MI | The team and the crowd cheered after the goal. |
Human | The crowd cheered when their team scored a goal. |
Concept Set | {bag, put, apple, tree, pick} |
T5-base-BL | A man puts a bag of apples on a tree. |
T5-base-KW | A man puts a bag under a tree and picks an apple. |
T5-base-Att | A man puts a bag under a tree and picks an apple. |
T5-base-P2T | A man puts a bag of apples on a tree and picks them. |
BART-base-MI | A man puts a bag of apple juice on a tree to pick it up |
Human | I picked an apple from the tree and put it in my bag. |
Concept Set | {circle, ball, throw, turn, hold} |
T5-large-BL | Someone turns and throws a ball in a circle. |
T5-large-KW | A man holds a ball and turns to throw it into a circle. |
T5-large-Att | A man holds a ball in a circle and throws it. |
T5-large-P2T | A man holds a ball, turns and throws it into a circle. |
BART-large-MI | He turns and throws a ball into the circle to hold it. |
Human | A girl holds the ball tightly, then turns to the left and throws the ball into the net which is in the shape of a circle. |
Concept Set | {hair, sink, lay, wash} |
T5-large-BL | A woman is washing her hair in a sink. |
T5-large-KW | A woman lays down to wash her hair in a sink. |
T5-large-Att | A man lays down to wash his hair in a sink. |
T5-large-P2T | A woman is washing her hair in a sink. |
BART-large-MI | A woman is washing her hair in the sink. She lay the sink down |
Human | The woman laid back in the salon chair, letting the hairdresser wash her hair in the sink. |
Concept Set | {wash, dry, towel, face} |
T5-large-BL | A man is washing his face with a towel. |
T5-large-KW | A man washes his face with a towel and then dries it. |
T5-large-Att | A man is washing his face with a towel and drying it. |
T5-large-P2T | A man is washing his face with a towel and drying it off. |
BART-large-MI | A man is washing his face with a towel to dry it. |
Human | The woman will wash the baby’s face and dry it with a towel. |
Model\Metrics | ROUGE-2/L | BLEU-3/4 | METEOR | CIDEr | SPICE | BERTScore | Cov | ||
Reported BART-large | 22.13 | 43.02 | 37.00 | 27.50 | 31.00 | 14.12 | 30.00 | - | 97.56 |
Reported T5-base | 15.33 | 36.20 | 28.10 | 18.00 | 24.60 | 9.73 | 23.40 | - | 83.77 |
Reported T5-Large | 21.98 | 44.41 | 40.80 | 30.60 | 31.00 | 15.84 | 31.80 | - | 97.04 |
Our BART-base | 15.91 | 36.15 | 38.30 | 28.30 | 30.20 | 15.07 | 30.35 | 58.26 | 93.44 |
Our BART-large | 17.27 | 37.32 | 39.95 | 30.20 | 31.15 | 15.72 | 31.20 | 58.58 | 95.03 |
Our T5-base | 17.27 | 37.69 | 41.15 | 31.00 | 31.10 | 16.37 | 32.05 | 60.32 | 94.44 |
Our T5-large | 17.90 | 38.31 | 43.80 | 33.60 | 32.70 | 17.02 | 33.45 | 61.39 | 96.26 |
Appendix C Human Evaluation Details
Human evaluation was done via paid crowdworkers on AMT, who were from Anglophone countries with lifetime approval rates % . Each example was evaluated by 2 annotators. The time given for each AMT task instance or HIT was 8 minutes. Sufficient time to read instructions, as calibrated by authors, was also considered in the maximum time limit for performing each HIT. Annotators were paid 98 cents per HIT. This rate (7.35$/hr) exceeds the minimum wage for the USA (7.2$/hr) and constitutes fair pay. We neither solicit, record, request, or predict any personal information pertaining to the AMT crowdworkers. Specific instructions and a question snippet can be seen in Figure 1.


Appendix D Graphs Displaying Other Hyperparameter Results
Figures 2, 3, 4, and 5 contain graphs displaying other hyperparameter results for Kw-aug, Att-aug, P2T, and Mask Infilling (MI), respectively.












Appendix E Further Qualitative Examples
See Table 15 for further qualitative examples.