This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Improving Prediction Backward-Compatiblility in NLP Model Upgrade with Gated Fusion

Yi-An Lai Elman Mansimov Yuqing Xie♡,  Yi Zhang
AWS AI Labs  University of Waterloo
{yianl,mansimov,yizhngn}@amazon.com
yuqing.xie@uwaterloo.ca
  Work done during author’s internship at AWS AI Labs.
Abstract

When upgrading neural models to a newer version, new errors that were not encountered in the legacy version can be introduced, known as regression111Within this work, regression denotes performance degradation in software systems, instead of the statistical technique for estimating relationships among variables.errors. This inconsistent behavior during model upgrade often outweighs the benefits of accuracy gain and hinders the adoption of new models. To mitigate regression errors from model upgrade, distillation and ensemble have proven to be viable solutions without significant compromise in performance. Despite the progress, these approaches attained an incremental reduction in regression which is still far from achieving backward-compatible model upgrade. In this work, we propose a novel method, Gated Fusion, that promotes backward compatibility via learning to mix predictions between old and new models. Empirical results on two distinct model upgrade scenarios show that our method reduces the number of regression errors by 62% on average, outperforming the strongest baseline by an average of 25%.

1 Introduction

Refer to caption
Figure 1: Illustration of regression errors when upgrading from BERT (Devlin et al., 2019) to ELECTRA (Clark et al., 2020) for classification. Red circles and green squares denote examples of different classes. Dashed lines represent decision boundaries.

In order to achieve a smooth continuous improvement of NLP applications, it is critical to guarantee consistent operation of the system after an upgrade. New errors introduced during the model upgrade interfere with the existing user experience and are considered to be a regression in the quality. Due to the difficulty of modularizing or explaining the behavior of deep neural networks, traditional software regression tests are inapplicable to neural based systems. The cost of arduous error analysis and model patching often exceeds the benefits of model upgrades. Developing methods that ensure backward compatibility during model upgrades without compromise in performance becomes a valuable research direction (Yan et al., 2021; Xie et al., 2021; Träuble et al., 2021; Cai et al., 2022).

The prediction backward-compatible model upgrade problem aims to improve consistency of correct classification predictions between legacy and upgraded models without accuracy loss. Yan et al. (2021) first studied backward compatibility during model upgrade on image classification tasks. They proposed to enforce the positive congruence of the new model with the old one by applying a knowledge distillation objective (Hinton et al., 2015) objective with re-weighting of training samples. Later, Xie et al. (2021) extended the work of Yan et al. (2021) by investigating the backward compatibility in NLP classification tasks. They found that their proposed distillation-based approach can help decrease the regression errors of specific linguistic phenomena in NLP classification tasks.

Despite progress with both distillation- and ensemble-based regression-mitigation approaches, there are limitations that prevent them from broad practical adoption in ML operations. Distillation-based methods attempt to transfer the prediction power of the old model to the new one on potential regression instances (Hinton et al., 2015). However, given the huge complexity of current neural architectures and relatively scarce training data in downstream tasks, models could have insufficient data to reliably estimate the probable regression cases and carry out the transfer on them (Xie et al., 2021; Cai et al., 2022). On the other hand, model ensemble aggregates predictions from differently-trained new models but bears no connection with the legacy version (Yan et al., 2021). These limitations reveal the two major challenges when striving to ensure backward compatibility. First, the new model could have distinct inductive bias and prediction behavior than the old system, rooted from inherent differences such as architecture, model size, and pretraining procedure (Liu et al., 2021). Second, during new model training, a reliable mechanism is needed in place to bridge the gap between two models and mitigate potential inconsistencies.

Inspired by the strength and weakness of prior approaches, we propose Gated Fusion to integrate old and new models via gating mechanism (Hochreiter and Schmidhuber, 1997; Chung et al., 2014; Gu et al., 2016), essentially a light-weight ensemble of legacy and upgrade models connected via a learned fusion gate. Specifically, we add a learned gate on top of the new model. We combine the logits from old and new models according to the weight from the gate. We train our Gated Fusion model by minimizing the standard cross-entropy error. The intuition is that the gate could learn to put more weights on the old model when the new model cannot produce correct predictions, effectively doing fall-backs that optimizes backward compatibility.

Empirical results demonstrate that our proposed approach outperforms other competing methods significantly, where we can obtain on average 62%62\% reduction of total negative flips, i.e. new errors caused by the model upgrade, without any degradation in accuracy performance. The effectiveness of Gated Fusion is validated across three diverse classification tasks and two distinct model upgrade scenarios (a) upgrade to larger model size (b) upgrade to advanced pretrained model, where consistent results are attained across the board.

Our main contributions are as follows:

  • We propose Gated Fusion that integrates old and new models via gating mechanism for backward-compatible model upgrade;

  • We evaluate competing methods on two distinct and challenging model upgrade scenarios across three diverse classification tasks;

  • Empirical results show that our proposed approach significantly outperforms competing methods and achieves regression reductions by a large margin across the board.

2 The Backward-Compatible Model Upgrade Problem

Refer to caption
Figure 2: Methods to improve prediction backward compatibility during model upgrade. (a) Distillation-based approach to align predicted logits on potential regression instances (Xie et al., 2021). (b) Ensemble of old and new models via weighted sum of either predicted logits or probabilities. (c) Our proposed Gated Fusion that learns a gate as a soft switch to dynamically determine whether to fall back to previous predictions.

The goal of backward-compatible model upgrade is to minimize regression errors without compromising the accuracy performance during model upgrade (Yan et al., 2021; Xie et al., 2021). In this work, we aim to improve the backward compatibility of model predictions in the NLP classification tasks. Following Xie et al. (2021), we study the scenario where the underlying pretrained language model (LM) is being upgraded.

Let xx be a natural language input with a class label y{1,2,,C}y\in\{1,2,...,C\}. 𝒟={xi,yi}i=1N\mathcal{D}=\{x_{i},y_{i}\}_{i=1}^{N} denotes a set of NN examples with corresponding labels. A classifier ff estimates the class probabilities given the input f(x)=(p(y=1|x),,p(y=C|x))\vec{f}(x)=(p(y=1|x),...,p(y=C|x))^{\top}. When upgrading from an old model foldf_{old} to a new model fnewf_{new}, normally with distinct architectures and trained on the same data, an improved model ff^{*} is produced based on foldf_{old} and fnewf_{new}. Our goal is for ff^{*} to minimize regression errors as an additional objective, while still achieving comparable performance to fnewof^{o}_{new}, the new model trained in the vanilla setting. Note that ff^{*} could be multiple times larger than fnewof^{o}_{new}, with model ensemble of fnewof^{o}_{new} as one example (Yan et al., 2021).

Measuring Backward Compatibility.

The backward compatibility is measured via quantifying regression errors on a given regression measurement set 𝒟reg={xi,yi}i=1M\mathcal{D}_{reg}=\{x_{i},y_{i}\}_{i=1}^{M}. 𝒟reg\mathcal{D}_{reg} could be a hidden customer test set comprising critical use cases, a set of behavioral testing examples for targeted evaluation (Ribeiro et al., 2020), or the development split from the dataset of interest. In this work, we take the development set as our 𝒟reg\mathcal{D}_{reg} for evaluation.

For classification, regression errors are characterized by negative flips, denoted as NF\mathcal{R}_{NF} – the portion of samples in 𝒟reg\mathcal{D}_{reg} that flip from correct prediction fold(xi)=yif_{old}(x_{i})=y_{i} to incorrect output fnew(xi)yif_{new}(x_{i})\neq y_{i} during model upgrade:

NF(𝒟reg,fold,fnew)=|{x|fold=y,fnewy}||𝒟reg|.\mathcal{R}_{NF}(\mathcal{D}_{reg},\vec{f}_{old},\vec{f}_{new})=\frac{|\{x|f_{old}=y,f_{new}\neq y\}|}{|\mathcal{D}_{reg}|}. (1)

One thing to emphasize is that maximizing classifier performance does not necessarily help in minimizing NF\mathcal{R}_{NF} (Yan et al., 2021; Xie et al., 2021).

3 Gated Fusion: Methodology

3.1 Method Overview

To improve backward compatibility in model upgrade, it’s crucial to have a mechanism that detects potential regression errors and mitigates them when making predictions. We propose Gated Fusion (GF) to achieve this by learning a gate as a soft switch to choose between generating predictions by the new model or resorting to outputs of the old model. Gated Fusion is inspired by the gating mechanism widely used in other applications. For example, mixing word copying mode with word generation mode for language modeling (Merity et al., 2016) and summarization (See et al., 2017).

Our proposed Gated Fusion fGFf^{*}_{GF} consists of the old model foldf_{old}, the new model fnewf_{new}, and a gating network gθg_{\theta}. The old model foldf_{old} is the legacy version before upgrade where the parameters are fixed. The new model fnewf_{new} has the same architecture as fnewof^{o}_{new} and is randomly initialized. The gating network gθg_{\theta} is a multi-layer feed-forward network with sigmoid function. It produces a scalar weight αgate\alpha_{gate} in the range [0,1][0,1] from the output layer of fnewf_{new}, denoted as new\mathcal{E}_{new}:

αgate(x)=gθ(new(x)).\alpha_{gate}(x)=g_{\theta}(\mathcal{E}_{new}(x)). (2)

We use αgate\alpha_{gate} to combine the logits of old and new models as our final outputs:

lGF(y|x)=(1αgate)lold(y|x)T+αgatelnew(y|x),l^{*}_{GF}(y|x)=(1-\alpha_{gate})\cdot\frac{l_{old}(y|x)}{T}+\alpha_{gate}\cdot l_{new}(y|x), (3)

where l(y|x)l(y|x) denotes predicted logits from models and TT is the temperature scaling to regularize the magnitude of old model’s logits. fnewf_{new} and gθg_{\theta} are then jointly trained end-to-end with cross-entropy loss between our output logits lGF(y|x)l^{*}_{GF}(y|x) and label distributions on downstream tasks.

The intuition behind Gated Fusion is that when fnewf_{new} makes a mistake while foldf_{old} produces the correct output, the gate gθg_{\theta} will learn to put more weight on foldf_{old} in order to minimize the final classification loss. This process effectively mitigates potential negative flips introduced by the model upgrade and thus improves the backward compatibility of final predictions.

3.2 Training and Inference

In practice, training Gated Fusion with randomly initialized fnewf_{new} would make the shallow gating network quickly converge to favor the fully-trained foldf_{old}. To prevent this, we only train fnewf_{new} for the first few epochs to ensure its competence before jointly training gθg_{\theta} and fnewf_{new} using lGF(x)l^{*}_{GF}(x). In addition, we found that stopping gradient flow from gθg_{\theta} to fnewf_{new} can further prevent the performance decrease of the new model within Gated Fusion:

αgate(x)=gθ(𝑠𝑡𝑜𝑝_𝑔𝑟𝑎𝑑(new(x))).\alpha_{gate}(x)=g_{\theta}(\mathit{stop\_grad}(\mathcal{E}_{new}(x))). (4)

At inference time, Gated Fusion produces logits from foldf_{old} and fnewf_{new} as well as the gate value αgate\alpha_{gate} to make output predictions:

fGF(x)=𝑆𝑜𝑓𝑡𝑚𝑎𝑥((1αgate)loldT+αgatelnew).f^{*}_{GF}(x)=\mathit{Softmax}\Big{(}(1-\alpha_{gate})\cdot\frac{l_{old}}{T}+\alpha_{gate}\cdot l_{new}\Big{)}. (5)

3.3 Inference with Cache

Our proposed Gated Fusion requires foldf_{old} to be hosted together with the new model. In reality, one could have a resource-constrained setting and request the old model to be discarded at inference. We note that in real applications, repetitive inputs are commonly seen in live traffic (Batrinca and Treleaven, 2015) and the backward compatibility of model upgrade entails that correct predictions can be preserved on the legacy instances already seen and predicted by the old model.

To simulate real scenarios, we randomly cache old model’s logits on a portion of test inputs. When getting out-of-cache instances, we use new model’s output embedding new(x)\mathcal{E}_{new}(x) as key and euclidean distance as metric to search for the nearest cached instance. The cached old-model logits can then be used for Gated Fusion to make predictions without hosting foldf_{old} at inference.

4 Experiments Setup

4.1 Model Upgrade Scenarios

We conduct experiments on two representative model upgrade scenarios: (a) upgrade to a larger pretrained model of the same type, where we use BERTbase to BERTlarge. (b) upgrade to a distinct pretrained model with the same size. We use BERTbase to ELECTRAbase (Clark et al., 2020) as this challenging model upgrade scenario for backward-compatibility, as they are pretrained under different self-supervised learning paradigms. The former uses masked language modeling (MLM) with reconstruction loss, while the latter is pretrained in generative-contrastive (adversarial) fashion with distributional divergence as the loss (Liu et al., 2021).

4.2 Datasets and Implementation

We evaluate our approach across three datasets. They represent different sentence-level classification tasks, from single-sentence to sentence-pair classification, with varying dataset sizes. We use: (a) Stanford Sentiment Treebank (SST-2), a single-sentence task to classify movie review sentiment, with 6767k train and 0.90.9k dev set (Socher et al., 2013). (b) Microsoft Research Paraphrase Corpus (MRPC) (Dolan and Brockett, 2005), a sentence-pair classification task for identifying paraphrases, with 3.73.7k train and 0.40.4k dev set. (c) Question Natural Language Inference (QNLI), a question-paragraph pair task to determine whether the paragraph contains the answer to the question, with 100100k train and 5.55.5k dev set. Datasets are taken from GLUE Benchmark (Wang et al., 2018) and processed with scripts from Hugging Face222https://huggingface.co/datasets/glue.

For implementation, we use the sequence classification and pre-trained model parameters from Hugging Face Transformers333https://huggingface.co/docs/transformers/index. Experiments are done in PyTorch (Paszke et al., 2019) with Tesla V100 GPUs and results are averaged over 55 random seeds. Learning rate, batch size, and train epoch are tuned during training new model alone on given tasks and then fixed for all backward-compatible solutions. In Gated Fusion, we first train fnewf_{new} alone for first (N1)(N-1) epochs and then jointly train gθg_{\theta} and fnewf_{new} with Gated Fusion logits lGFl^{*}_{GF} in the last epoch. Further implementation details can be found in the Appendix.

4.3 Baselines

We compare our approach with several strong baselines. (a) Train the new model directly on the target task without any adjustment, i.e. fnewof^{o}_{new}. (b) The specialized distillation method proposed in Xie et al. (2021), where the KL-divergence of prediction probabilities between old and new models is applied when pold(y=yi|xi)>pnew(y=yi|xi)p_{old}(y=y_{i}|x_{i})>p_{new}(y=y_{i}|x_{i}). (c) Model ensemble via majority-voting that was shown to be very effective (Yan et al., 2021; Xie et al., 2021). Similarly, we use 55-seed new model ensemble as a strong baseline. (d) The ensemble of the old and new models probabilities, p(y|x)=(1α)pold(y|x)+αpnew(y|x)p^{*}(y|x)=(1-\alpha)\cdot p_{old}(y|x)+\alpha\cdot p_{new}(y|x), as well as ensemble of the old and new models logits, l(y|x)=(1α)lold(y|x)+αlnew(y|x)l^{*}(y|x)=(1-\alpha)\cdot l_{old}(y|x)+\alpha\cdot l_{new}(y|x), where α\alpha is searched among [0.5,0.6,0.7,0.8,0.9][0.5,0.6,0.7,0.8,0.9] to maximize backward-compatibility while achieving accuracy on par with the vanilla fnewof^{o}_{new}.

5 Results and Analysis

SST-2 MRPC QNLI
BERTbase \rightarrow BERTlarge NF\mathcal{R}_{NF} Accuracy NF\mathcal{R}_{NF} Accuracy NF\mathcal{R}_{NF} Accuracy
Old Model - 92.000.27 - 85.690.90 - 90.740.09
New Model 2.180.21 93.120.29 4.121.04 87.401.02 2.720.13 92.220.16
Distillation (Xie et al., 2021) 1.970.22 93.330.20 3.530.77 87.701.34 2.310.14 92.600.19
New Model Ensemble 2.000.31 93.300.24 2.250.61 88.870.77 1.980.21 92.970.22
Old-New Probs Ensemble 1.060.27 93.120.38 1.670.78 87.161.12 1.040.26 92.440.23
Old-New Logits Ensemble 1.060.27 93.120.38 1.670.78 87.161.12 1.040.26 92.440.23
Gated Fusion 0.780.20 93.050.09 1.180.52 87.450.52 0.730.13 92.240.24
Table 1: Negative flip rate NF\mathcal{R}_{NF} and model accuracy (%\%) of competing methods to optimize backward-compatibility without performance degradation during BERTbase \rightarrow BERTlarge model upgrade.
SST-2 MRPC QNLI
BERTbase \rightarrow ELECTRAbase NF\mathcal{R}_{NF} Accuracy NF\mathcal{R}_{NF} Accuracy NF\mathcal{R}_{NF} Accuracy
Old Model - 92.000.27 - 85.690.90 - 90.740.09
New Model 1.630.20 95.000.06 3.730.36 88.580.57 2.820.32 92.900.26
Distillation (Xie et al., 2021) 1.490.24 95.020.21 3.680.79 88.820.94 2.580.17 93.030.16
New Model Ensemble 1.120.09 95.390.09 3.240.24 89.020.48 2.260.08 93.490.07
Old-New Probs Ensemble 1.400.17 95.070.15 3.140.42 88.530.48 0.980.20 93.040.21
Old-New Logits Ensemble 0.890.17 94.950.13 3.280.43 88.480.51 0.980.20 93.040.21
Gated Fusion 0.710.18 95.020.16 2.400.50 88.680.68 0.810.16 92.980.17
Table 2: Negative flip rate NF\mathcal{R}_{NF} and model accuracy (%\%) of competing methods to optimize backward-compatibility without performance degradation during BERTbase \rightarrow ELECTRAbase model upgrade.

5.1 Upgrade to a Larger Pretrained Model

Our first model upgrade scenario scales up the size of underlying pretrained language models. We experiment with BERTbase to BERTlarge, where the model size is tripled (110M vs 340M) and the model depth is doubled (12 vs 24 layers).

Table 1 shows the results. For fnewof^{o}_{new}, we can observe that the negative flip rates NF\mathcal{R}_{NF} are usually much larger than the accuracy gains across tasks, which could be the reason to hinder new model adoptions in real-world applications. Besides, when dividing NF\mathcal{R}_{NF} over the error rate (1accuracy)(1-\textit{accuracy}), we can observe that around 30%30\% to 40%40\% of all fnewof^{o}_{new} prediction errors are in fact the new errors introduced during model upgrade. For improving prediction backward-compatibility, our proposed Gated Fusion outperforms other competing methods to considerably reduce NF\mathcal{R}_{NF} without degradation on accuracy. Note that best α\alpha values found for the two variants of old-new ensemble are both 0.50.5, hence producing identical results.

Compared to the vanilla new model, gated fusion obtains absolute NF\mathcal{R}_{NF} reductions of −1.40%1.40\% on SST-2, −2.94%2.94\% on MRPC, and −1.99%1.99\% on QNLI. These translate to reducing the total negative flip cases by 64.2%64.2\%, 71.4%71.4\%, 73.2%73.2\%, respectively. Compared to the strongest baseline (old-new ensemble), we obtain further absolute NF\mathcal{R}_{NF} reductions of −0.28%0.28\% on SST-2, −0.49%0.49\% on MRPC, and −0.31%0.31\% on QNLI, which translate to further reducing 12.8%12.8\%, 11.9%11.9\%, and 11.4%11.4\% of negative flip cases. These results show the effectiveness of our method to mitigate a significant amount of regression errors during model upgrade.

SST-2 MRPC QNLI
NF\mathcal{R}_{NF} Accuracy NF\mathcal{R}_{NF} Accuracy NF\mathcal{R}_{NF} Accuracy
Old Model: BERTbase - 92.000.27 - 85.690.90 - 90.740.09
New Model: BERTlarge 2.180.21 93.120.29 4.121.04 87.401.02 2.720.13 92.220.16
Model Ensemble: 5 seeds 2.000.31 93.300.24 2.250.61 88.870.77 1.980.21 92.970.22
Model Ensemble: 10 seeds 1.790.17 93.690.15 2.010.29 89.460.51 2.010.14 92.970.19
Model Ensemble: 20 seeds 1.790.25 93.620.16 1.760.50 89.560.48 1.820.15 93.130.08
Gated Fusion 0.780.20 93.050.09 1.180.52 87.450.52 0.730.13 92.240.24
New Model: ELECTRAbase 1.630.20 95.000.06 3.730.36 88.580.57 2.820.32 92.900.26
Model Ensemble: 5 seeds 1.120.09 95.390.09 3.240.24 89.020.48 2.260.08 93.490.07
Model Ensemble: 10 seeds 1.240.18 95.300.16 3.630.50 88.580.20 2.210.12 93.570.15
Model Ensemble: 20 seeds 1.190.16 95.320.17 3.430.51 88.920.48 2.150.17 93.630.11
Gated Fusion 0.710.18 95.020.16 2.400.50 88.680.68 0.810.16 92.980.17
Table 3: Negative flip rate NF\mathcal{R}_{NF} and model accuracy (%\%) when increasing number of seeds used in new model ensemble, comparing with our proposed method (Gated Fusion).

5.2 Upgrade to a Different Pretrained Model

A more challenging upgrade scenario is when old and new models are pretrained under distinctive paradigms, producing two representation spaces of fairly different characteristics (Meng et al., 2021b). We experiment with BERTbase to ELECTRAbase in this scenario, where two models have the same size but are pretrained under utterly different schemes, i.e. generative versus adversarial.

Table 2 shows the results. For fnewof^{o}_{new}, compared with upgrading to BERTlarge, we observe larger accuracy gains and lower NF\mathcal{R}_{NF} on SST-2 and MRPC. However, on QNLI, upgrading to ELECTRAbase achieves a higher accuracy gain but an even a higher NF\mathcal{R}_{NF}. This implies that boosting accuracy and improving backward compatibility could be two related but different objectives.

For mitigation strategies, Gated Fusion achieves the lowest negative flip rates across datasets without any accuracy loss. We obtain absolute NF\mathcal{R}_{NF} reductions of −0.92%0.92\% on SST-2, −1.33%1.33\% on MRPC, and −2.01%2.01\% on QNLI over the vanilla setup, reducing 56.4%56.4\%, 35.7%35.7\%, and 71.3%71.3\% of overall negative flips, respectively. Compared with upgrading to BERTlarge, we observe that upgrading to ELECTRAbase has much smaller relative negative flip reductions on SST-2 and MRPC, showing that it could be indeed harder to improve backward-compatibility when upgrading to a distinct pretrained model. In contrast, similar relative negative flip reductions are observed on QNLI across two upgrade scenarios. This could be attributed to the abundant training data of the downstream task.

SST-2 MRPC QNLI
Old: BERTbase 92.000.27 85.690.90 90.740.09
to BERTlarge 93.120.29 87.401.02 92.220.16
Gated Fusion 93.050.09 87.450.52 92.240.24
     - drop old model 93.170.61 87.751.14 92.220.44
to ELECTRAbase 95.000.06 88.580.57 92.900.26
Gated Fusion 95.020.16 88.680.68 92.980.17
     - drop old model 95.160.09 88.630.94 93.060.13
Table 4: Accuracy (%\%) when dropping the old model within Gated Fusion at inference time.
SST-2
NF\mathcal{R}_{NF} Accuracy
Old Model: BERTbase - 92.000.27
New Model: ELECTRAbase 1.630.20 95.000.06
Gated Fusion - 50%\% cache 1.260.10 94.860.27
Gated Fusion - 75%\% cache 0.990.25 94.910.12
Gated Fusion 0.710.18 95.020.16
Table 5: Negative flip rate NF\mathcal{R}_{NF} and model accuracy (%\%) of Gated Fusion with X%X\% cache of old model logits at inference time.

5.3 Drop Old Model at Inference Time

Our proposed method requires the old model to be hosted together with the new model. A natural question is whether we could train Gated Fusion with the old model and then discard it at inference time to host the new model only.

We first experiment with directly dropping the old model within Gated Fusion at inference time. Results in Table 4 show that dropping old model in Gated Fusion can still achieve comparable accuracy across the board, suggesting no performance degradation. Nonetheless, we observe that the negative flip rates also fall back to similar positions as training the new model in the vanilla setting.

However, in real application scenario, live inputs are often repetitively seen across time and ensuring backward-compatibility means that correct predictions on same instances can be preserved after model upgrade. We experiment with the caching method introduced in section 3.3 to store old model’s logits on random X%X\% of test instances where Gated Fusion can later access them for inference. Results in Table 5 show that with higher percentage of cache, NF\mathcal{R}_{NF} is gradually reduced towards NF\mathcal{R}_{NF} of the original Gated Fusion, which is equivalent to 100%100\% cache. Still, we observe a notable gap in NF\mathcal{R}_{NF} between the partial caching and full settings. We leave the examination of ways to achieve the upper bound in reduction in NF\mathcal{R}_{NF} with smaller cache to the future work.

(Task, Label) Examples
BERTbase \rightarrow BERTlarge (SST-2, Positive) [Sentence] A study in shades of gray, offering itself up in subtle plot maneuvers …
(SST-2, Negative) [Sentence] Manages to be both repulsively sadistic and mundane.
(MRPC, Not Equivalent) [Sentence 1] Vivace was founded in 1999 and has raised over $118 million in three rounds of venture financing. [Sentence 2] During difficult times for technology venture capital, Vivace raised over $118 million in three rounds of venture financing.
(QNLI, Entailment) [Question] Why was there a depreciation of the industrialized nations dollars? [Sentence] Anticipating that currency values would fluctuate unpredictably for a time, the industrialized nations increased their reserves (by expanding their money supplies) in amounts far greater than before.
BERTbase \rightarrow ELECTRAbase (SST-2, Positive) [Sentence] Aside from minor tinkering , this is the same movie you probably loved in 1994, except that it looks even better.
(SST-2, Negative) [Sentence] It showcases carvey’s talent for voices, but not nearly enough and not without taxing every drop of one’s patience to get to the good stuff .
(MRPC, Equivalent) [Sentence 1] Blair’s Foreign Secretary Jack Straw was to take his place on Monday to give a statement to parliament on the European Union. [Sentence 2] Blair’s office said his Foreign Secretary Jack Straw would take his place on Monday to give a statement to parliament on the EU meeting the prime minister attended last week.
(QNLI, Not Entailment) [Question] What is the main executive body of the EU? [Sentence] This means that the Commission has a monopoly on initiating the legislative procedure, although the Council is the "de facto catalyst of many legislative initiatives".
Table 6: Examples of regression errors present when upgrading to the vanilla new model fnewof^{o}_{new} but fixed by our Gated Fusion approach, i.e. predictions of (fold,fnewo,fGF)(f_{old},f^{o}_{new},f^{*}_{GF}) are (correct, incorrect, correct), respectively.

5.4 Limitations of New Model Ensemble

In previous works (Yan et al., 2021; Xie et al., 2021), new model ensemble via majority voting is shown to effectively reduce negative flips and posed as a difficult-to-beat baseline. Here, we increase the number of models in ensemble to examine its limitations. Results in Table 3 show that ensemble with more models generally help to obtain lower NF\mathcal{R}_{NF}. However, NF\mathcal{R}_{NF} converges quickly as number of models increased, where a notable gap remains between new model ensemble and Gated Fusion. Moreover, the results show once more that boosting accuracy does not necessarily improve the backward compatibility in model upgrade.

In principle, two sources could cause negative flips during model upgrade (a) the stochasticity during model training, including initializations, data loading order, and optimization process (Somepalli et al., 2022). (b) the distinctions between old and new model hypotheses, including architecture and pretraining data and procedure, leading to different representation space structures and prediction behaviors in terms of decision boundaries. Without an explicit connection to foldf_{old}, new model ensemble can only reduce negative flips primarily caused by the first factor, while our proposed Gated Fusion directly learns to mitigate regression errors regardless of their causes.

Besides, as large-scale generative models become more and more powerful and popular (Raffel et al., 2020; Brown et al., 2020; Su et al., 2021), it would be difficult to fine-tune them multiple times on a target task for ensemble.

5.5 Analysis of Gated Fusion

Comparing fnewof^{o}_{new} with fGFf^{*}_{GF}, we can calculate the fix rate and new fault rate of our Gated Fusion method. During an upgrade, if there are 2020 negative flips with fnewof^{o}_{new} and 1616 out of them can be mitigated by fGFf^{*}_{GF}, we obtain the fix rate to be 16/20=80%16/20=80\%. Similarly, if fGFf^{*}_{GF} introduces another 44 new negative flips which are not present with fnewof^{o}_{new}, the new fault rate is computed to be 4/20=20%4/20=20\%. We calculate the 55-seed average of these two rates across different classification tasks and upgrade scenarios. In BERTbase to BERTlarge, the averaged fix rates by Gated Fusion are 68.4%68.4\% on SST-2, 83.8%83.8\% on MRPC, and 82.9%82.9\% on QNLI, with new fault rates being 4.1%4.1\% on SST-2, 11.3%11.3\% on MRPC, and 9.7%9.7\% on QNLI. In BERTbase to ELECTRAbase, Gated Fusion achieves the averaged fix rates 58.0%58.0\% on SST-2, 50.8%50.8\% on MRPC, and 75.6%75.6\% on QNLI, with new fault rates being 2.8%2.8\% on SST-2, 15.2%15.2\% on MRPC, and 4.0%4.0\% on QNLI. These results show that, on average, Gated Fusion is able to eliminate 69.9%69.9\% of total regression errors while adding only 7.9%7.9\% new ones, comparing with doing model upgrade without any treatment, i.e. fnewof^{o}_{new}.

Table 6 shows a few regression error cases fixed by our proposed approach. In general, Gated Fusion can mitigate negative flips happened on different classes across diverse tasks as well as on inputs with variable lengths. With closer inspections of fGFf^{*}_{GF}, we found that when fnewf_{new} produces incorrect predictions and foldf_{old} gives correct outputs, gθg_{\theta} is capable of putting larger weights on foldf_{old} to ensure the backward compatibility. We also observed that the gate gθg_{\theta} is more prone to over-fitting when the downstream tasks have smaller training set, e.g. MRPC, or are more difficult in nature, e.g. single-sentence task SST-2 versus sentence-pair tasks, which causes Gated Fusion to introduce more new errors, i.e. higher new fault rates.

6 Discussion

Gated Fusion requires to host both old and new models at inference time, which could raise a concern regarding the increased computational burden. However, in practice, old model’s logits of previous inference instances can be cached in storage and later leveraged in our Gated Fusion. That is, we only need to host the new model with the gate at inference time and leverage old predictions from cache. And for the out-of-cache inputs, backward-compatibility would be less of an issue since users have not observed such examples to make conclusions on the underlying regression.

For real-world applications, there could be multiple model updates and thus multiple legacy versions. We note that in this scenario, user experience would be primarily grounded on predictions of the latest legacy version, which are also saved in cache. Our Gated Fusion can hence leverage them and make new model’s predictions compatible to those from the latest legacy version.

In addition, we emphasize that the main challenge in the regression reduction research problem is to find the best trade-off between model effectiveness and backward compatibility. In this work, we show that the weighted ensemble of old-new models with a learned gate, which we call Gated Fusion, achieves a better negative flip rate than previously explored methods for regression reduction, while straight-forward ensemble approaches cannot naturally weigh on this trade-off. We don’t claim to invent the gated ensemble of old and new models but rather that our main contribution is to show that by repurposing the classic gating mechanism, the gated ensemble can become the most competitive approach to the challenging model-upgrade regression reduction problem, with no overall performance degradation on two realistic model update scenarios across three different datasets.

Recently, more and more NLP products have been deployed in the industry as this field matures. We would like to stress that as better NLP models are being developed, the backward-compatible model upgrade problem naturally emerges as the new research topic strongly motivated by the real-world challenges. While backward-compatibility is currently a niche research topic, we believe that there are many thrilling future directions worth to be investigated.

7 Related Work

Yan et al. (2021) first studied the backward compatibility of predictions during model upgrade on image classification tasks. Later, Xie et al. (2021) investigated the similar topic in natural language understanding and formulated it as a constrained optimization problem. They both show that customized variants of knowledge distillation (Hinton et al., 2015), which align the predictions of old and new models on potential regression errors, are effective approaches. A model ensemble has also shown to be surprisingly effective (Yan et al., 2021; Xie et al., 2021), despite no explicit connection between old and new models. This was credited to variance reduction in model predictions, making it less prone to over-fitting and reducing regression errors indirectly. In this work, we leverage the gating mechanism to combine old and new models to further reduce model upgrade regression errors by a large margin across classification tasks.

Cai et al. (2022) analyzed and proposed backward congruent re-ranking to reduce regression in model upgrades for structured predictions tasks such as dependency parsing and conversational semantic parsing. Träuble et al. (2021) proposed an efficient probabilistic approach to locate data instances whose old predictions could be incorrect and update them with ones from the new model. Zhou et al. (2022) looked into forward compatibility, where new classes can be easily incorporated without negatively impacting existing prediction behavior. More recently, Schumann et al. (2023) inspected classification model regression during training data updates and mitigated the problem by interpolating between weights of the old and new models. On top of that, learning cross-model compatible embeddings has been extensively explored in visual search (Chen et al., 2019; Hu et al., 2019; Wang et al., 2020). Several techniques have been proposed to optimize cross-model interoperability of embeddings, including metric space alignment (Shen et al., 2020), architecture search (Duggal et al., 2021), and aligning class centers between models Meng et al. (2021a). In this work, we focus on improving backward compatibility during model upgrade in terms of prediction behavior on classification tasks, i.e. old and new models should produce consistently correct predictions.

Reducing regression during model upgrade is also related to continual learning (Parisi et al., 2019; De Lange et al., 2019; Sun et al., 2019; Chuang et al., 2020; Sachidananda et al., 2021), incremental learning (Chaudhry et al., 2018; Shan et al., 2020) and concept drifting (Gama et al., 2014; Žliobaitė et al., 2016; Ganin et al., 2016; Zhuang et al., 2020; Lazaridou et al., 2021). In these problems, models are required to learn from and deal with continuously changing data (in terms of examples, classes or tasks), and also need to prevent the forgetting of previously learnt knowledge. This could be one potential cause of regression observed at inference. However, in backward-compatible model upgrade, a new model, usually with distinct network architecture, is trained from scratch to perform the same task and is expected to behave similarly wherever the previous model predicts correctly.

The gating mechanism is widely adopted by recurrent neural networks to effectively control information flows across networks (Hochreiter and Schmidhuber, 1997; Cho et al., 2014; Van Oord et al., 2016; Dauphin et al., 2017; Lai et al., 2019) and contextualize embeddings Peters et al. (2018); Lai et al. (2020). It is then repurposed to act as a switch for the mixture of different prediction modes, notably to combine input word copying based on the pointer network (Vinyals et al., 2015) with the word generation from output vocabulary (Gu et al., 2016; Merity et al., 2016; See et al., 2017). Our proposed approach is inspired by these works and leverages the gating mechanism to effectively combine old and new models to improve backward compatibility during model upgrade.

8 Conclusion

Ensuring backward compatibility during model upgrade has become a critical topic in real-world NLP applications. In this work, we proposed a new approach, Gated Fusion, that achieves significantly better backward compatibility without compromising accuracy performance on two challenging upgrade scenarios for NLP classification. Experiments demonstrated that our approach outperforms competing methods and achieves negative flip rate reductions by up to 73.2%73.2\%. Our future research includes improving backward compatibility beyond classification to span detection, model upgrades with very large language models, and upgrades on training data or label schema. We hope that this work can inspire further research and make progress towards smoother transitions of prediction powers as NLP systems evolve.

Limitations

Our proposed method mostly works on the upgrades of underlying pretrained language models for NLP classification tasks. Potential limitations include applying our approach on distant tasks such as question answering or information retrieval, upgrade to models from different architecture families such as recurrent neural nets, and the inapplicability of our method to more recent learning formulation such as in-context learning via prompting.

Ethics Statement

Prediction backward compatibility during model upgrade is an emerging research topic to ensure positive congruency and smoother transitions from existing models towards more performant systems. With primary evaluation on accuracy and negative flips, we acknowledge that our method may also inherit social biases and other toxicity persisted in the legacy models. On the other hand, we have noted that fairness and safety have been one of principal criteria when developing system upgrades. Investigations of the inheritance of persistent toxicity and mitigation of it during backward-compatible upgrades merit interests of future research.

Acknowledgements

We would like to acknowledge AWS AI Labs for inspiring discussions, honest feedback, and full support. We are also very grateful to reviewers for judicious comments and valuable suggestions.

References

  • Batrinca and Treleaven (2015) Bogdan Batrinca and Philip C Treleaven. 2015. Social media analytics: a survey of techniques, tools and platforms. Ai & Society, 30(1):89–116.
  • Brown et al. (2020) Tom B Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. 2020. Language models are few-shot learners. arXiv preprint arXiv:2005.14165.
  • Cai et al. (2022) Deng Cai, Elman Mansimov, Yi-An Lai, Yixuan Su, Lei Shu, and Yi Zhang. 2022. Measuring and reducing model update regression in structured prediction for nlp. arXiv preprint arXiv:2202.02976.
  • Chaudhry et al. (2018) Arslan Chaudhry, Puneet K Dokania, Thalaiyasingam Ajanthan, and Philip HS Torr. 2018. Riemannian walk for incremental learning: Understanding forgetting and intransigence. In Proceedings of the European Conference on Computer Vision (ECCV), pages 532–547.
  • Chen et al. (2019) Ken Chen, Yichao Wu, Haoyu Qin, Ding Liang, Xuebo Liu, and Junjie Yan. 2019. R3 adversarial network for cross model face recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9868–9876.
  • Cho et al. (2014) Kyunghyun Cho, Bart Van Merriënboer, Dzmitry Bahdanau, and Yoshua Bengio. 2014. On the properties of neural machine translation: Encoder-decoder approaches. arXiv preprint arXiv:1409.1259.
  • Chuang et al. (2020) Yung-Sung Chuang, Shang-Yu Su, and Yun-Nung Chen. 2020. Lifelong language knowledge distillation. arXiv preprint arXiv:2010.02123.
  • Chung et al. (2014) Junyoung Chung, Caglar Gulcehre, KyungHyun Cho, and Yoshua Bengio. 2014. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555.
  • Clark et al. (2020) Kevin Clark, Minh-Thang Luong, Quoc V. Le, and Christopher D. Manning. 2020. Electra: Pre-training text encoders as discriminators rather than generators. In International Conference on Learning Representations.
  • Dauphin et al. (2017) Yann N Dauphin, Angela Fan, Michael Auli, and David Grangier. 2017. Language modeling with gated convolutional networks. In International conference on machine learning, pages 933–941. PMLR.
  • De Lange et al. (2019) Matthias De Lange, Rahaf Aljundi, Marc Masana, Sarah Parisot, Xu Jia, Ales Leonardis, Gregory Slabaugh, and Tinne Tuytelaars. 2019. Continual learning: A comparative study on how to defy forgetting in classification tasks. arXiv preprint arXiv:1909.08383, 2(6).
  • Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics.
  • Dolan and Brockett (2005) William B. Dolan and Chris Brockett. 2005. Automatically constructing a corpus of sentential paraphrases. In Proceedings of the Third International Workshop on Paraphrasing (IWP2005).
  • Duggal et al. (2021) Rahul Duggal, Hao Zhou, Shuo Yang, Yuanjun Xiong, Wei Xia, Zhuowen Tu, and Stefano Soatto. 2021. Compatibility-aware heterogeneous visual search.
  • Gama et al. (2014) João Gama, Indrė Žliobaitė, Albert Bifet, Mykola Pechenizkiy, and Abdelhamid Bouchachia. 2014. A survey on concept drift adaptation. ACM computing surveys (CSUR), 46(4):1–37.
  • Ganin et al. (2016) Yaroslav Ganin, Evgeniya Ustinova, Hana Ajakan, Pascal Germain, Hugo Larochelle, François Laviolette, Mario Marchand, and Victor Lempitsky. 2016. Domain-adversarial training of neural networks. The journal of machine learning research, 17(1):2096–2030.
  • Gu et al. (2016) Jiatao Gu, Zhengdong Lu, Hang Li, and Victor OK Li. 2016. Incorporating copying mechanism in sequence-to-sequence learning. arXiv preprint arXiv:1603.06393.
  • Hinton et al. (2015) Geoffrey E. Hinton, Oriol Vinyals, and J. Dean. 2015. Distilling the knowledge in a neural network. ArXiv, abs/1503.02531.
  • Hochreiter and Schmidhuber (1997) Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. Neural computation, 9(8):1735–1780.
  • Hu et al. (2019) Jie Hu, Rongrong Ji, Hong Liu, Shengchuan Zhang, Cheng Deng, and Qi Tian. 2019. Towards visual feature translation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3004–3013.
  • Lai et al. (2019) Yi-An Lai, Arshit Gupta, and Yi Zhang. 2019. Goal-embedded dual hierarchical model for task-oriented dialogue generation. arXiv preprint arXiv:1909.09220.
  • Lai et al. (2020) Yi-An Lai, Garima Lalwani, and Yi Zhang. 2020. Context analysis for pre-trained masked language models. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 3789–3804.
  • Lazaridou et al. (2021) Angeliki Lazaridou, Adhiguna Kuncoro, Elena Gribovskaya, Devang Agrawal, Adam Liska, Tayfun Terzi, Mai Gimenez, Cyprien de Masson d’Autume, Sebastian Ruder, Dani Yogatama, et al. 2021. Pitfalls of static language modelling. arXiv preprint arXiv:2102.01951.
  • Liu et al. (2021) Xiao Liu, Fanjin Zhang, Zhenyu Hou, Li Mian, Zhaoyu Wang, Jing Zhang, and Jie Tang. 2021. Self-supervised learning: Generative or contrastive. IEEE Transactions on Knowledge and Data Engineering.
  • Meng et al. (2021a) Qiang Meng, Chixiang Zhang, Xiaoqiang Xu, and Feng Zhou. 2021a. Learning compatible embeddings. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 9939–9948.
  • Meng et al. (2021b) Yu Meng, Chenyan Xiong, Payal Bajaj, Paul Bennett, Jiawei Han, Xia Song, et al. 2021b. Coco-lm: Correcting and contrasting text sequences for language model pretraining. Advances in Neural Information Processing Systems, 34:23102–23114.
  • Merity et al. (2016) Stephen Merity, Caiming Xiong, James Bradbury, and Richard Socher. 2016. Pointer sentinel mixture models. arXiv preprint arXiv:1609.07843.
  • Parisi et al. (2019) German I Parisi, Ronald Kemker, Jose L Part, Christopher Kanan, and Stefan Wermter. 2019. Continual lifelong learning with neural networks: A review. Neural Networks, 113:54–71.
  • Paszke et al. (2019) Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. 2019. Pytorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems 32, pages 8024–8035.
  • Peters et al. (2018) Matthew E Peters, Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark, Kenton Lee, and Luke Zettlemoyer. 2018. Deep contextualized word representations. arXiv preprint arXiv:1802.05365.
  • Raffel et al. (2020) Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J Liu. 2020. Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research, 21(1):5485–5551.
  • Ribeiro et al. (2020) Marco Tulio Ribeiro, Tongshuang Wu, Carlos Guestrin, and Sameer Singh. 2020. Beyond accuracy: Behavioral testing of NLP models with CheckList. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 4902–4912, Online. Association for Computational Linguistics.
  • Sachidananda et al. (2021) Vin Sachidananda, Jason S Kessler, and Yi-An Lai. 2021. Efficient domain adaptation of language models via adaptive tokenization. arXiv preprint arXiv:2109.07460.
  • Schumann et al. (2023) Raphael Schumann, Elman Mansimov, Yi-An Lai, Nikolaos Pappas, Xibin Gao, and Yi Zhang. 2023. Backward compatibility during data updates by weight interpolation. arXiv preprint arXiv:2301.10546.
  • See et al. (2017) Abigail See, Peter J Liu, and Christopher D Manning. 2017. Get to the point: Summarization with pointer-generator networks. arXiv preprint arXiv:1704.04368.
  • Shan et al. (2020) Guangxu Shan, Shiyao Xu, Li Yang, Shengbin Jia, and Yang Xiang. 2020. Learn#: A novel incremental learning method for text classification. Expert Systems with Applications, 147:113198.
  • Shen et al. (2020) Yantao Shen, Yuanjun Xiong, Wei Xia, and Stefano Soatto. 2020. Towards backward-compatible representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
  • Socher et al. (2013) Richard Socher, Alex Perelygin, Jean Wu, Jason Chuang, Christopher D. Manning, Andrew Ng, and Christopher Potts. 2013. Recursive deep models for semantic compositionality over a sentiment treebank. In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, pages 1631–1642.
  • Somepalli et al. (2022) Gowthami Somepalli, Liam Fowl, Arpit Bansal, Ping Yeh-Chiang, Yehuda Dar, Richard Baraniuk, Micah Goldblum, and Tom Goldstein. 2022. Can neural nets learn the same model twice? investigating reproducibility and double descent from the decision boundary perspective. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 13699–13708.
  • Su et al. (2021) Yixuan Su, Lei Shu, Elman Mansimov, Arshit Gupta, Deng Cai, Yi-An Lai, and Yi Zhang. 2021. Multi-task pre-training for plug-and-play task-oriented dialogue system. arXiv preprint arXiv:2109.14739.
  • Sun et al. (2019) Fan-Keng Sun, Cheng-Hao Ho, and Hung-Yi Lee. 2019. Lamol: Language modeling for lifelong language learning. In International Conference on Learning Representations.
  • Träuble et al. (2021) Frederik Träuble, Julius Von Kügelgen, Matthäus Kleindessner, Francesco Locatello, Bernhard Schölkopf, and Peter Gehler. 2021. Backward-compatible prediction updates: A probabilistic approach. Advances in Neural Information Processing Systems, 34:116–128.
  • Van Oord et al. (2016) Aaron Van Oord, Nal Kalchbrenner, and Koray Kavukcuoglu. 2016. Pixel recurrent neural networks. In International conference on machine learning, pages 1747–1756. PMLR.
  • Vinyals et al. (2015) Oriol Vinyals, Meire Fortunato, and Navdeep Jaitly. 2015. Pointer networks. Advances in neural information processing systems, 28.
  • Wang et al. (2018) Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel Bowman. 2018. GLUE: A multi-task benchmark and analysis platform for natural language understanding. In Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, pages 353–355.
  • Wang et al. (2020) Chien-Yi Wang, Ya-Liang Chang, Shang-Ta Yang, Dong Chen, and Shang-Hong Lai. 2020. Unified representation learning for cross model compatibility. arXiv preprint arXiv:2008.04821.
  • Xie et al. (2021) Yuqing Xie, Yi an Lai, Yuanjun Xiong, Yi Zhang, and Stefano Soatto. 2021. Regression bugs are in your model! measuring, reducing and analyzing regressions in nlp model updates. arXiv preprint arXiv:2105.03048.
  • Yan et al. (2021) Sijie Yan, Yuanjun Xiong, Kaustav Kundu, Shuo Yang, Siqi Deng, Meng Wang, Wei Xia, and Stefano Soatto. 2021. Positive-congruent training: Towards regression-free model updates. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 14299–14308.
  • Zhou et al. (2022) Da-Wei Zhou, Fu-Yun Wang, Han-Jia Ye, Liang Ma, Shiliang Pu, and De-Chuan Zhan. 2022. Forward compatible few-shot class-incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9046–9056.
  • Zhuang et al. (2020) Fuzhen Zhuang, Zhiyuan Qi, Keyu Duan, Dongbo Xi, Yongchun Zhu, Hengshu Zhu, Hui Xiong, and Qing He. 2020. A comprehensive survey on transfer learning. Proceedings of the IEEE, 109(1):43–76.
  • Žliobaitė et al. (2016) Indrė Žliobaitė, Mykola Pechenizkiy, and Joao Gama. 2016. An overview of concept drift applications. Big data analysis: new algorithms for a new society, pages 91–114.

Appendix A Details on Experiment Settings

A.1 Model Training Hyper-parameters

We search among following hyper-parameter space for the training of the old model foldf_{old} and the new model in the vanilla setting fnewof^{o}_{new} across all datasets:

  • Learning Rate: 5e65e^{-6}, 1e51e^{-5}, 3e53e^{-5}, 5e55e^{-5}

  • Batch Size: 1616, 3232

  • Training Epochs: 33, 55, 88.

The selected hyper-parameters for each model with (learning rate, batch size, training epoch):

  • BERTbase:

    • On SST-2: (lr 1e5,batch 16,epoch 5)(\text{lr }1e^{-5},\text{batch }16,\text{epoch }5)

    • On MRPC: (lr 3e5,batch 16,epoch 5)(\text{lr }3e^{-5},\text{batch }16,\text{epoch }5)

    • On QNLI: (lr 3e5,batch 32,epoch 3)(\text{lr }3e^{-5},\text{batch }32,\text{epoch }3)

  • BERTlarge:

    • On SST-2: (lr 1e5,batch 16,epoch 5)(\text{lr }1e^{-5},\text{batch }16,\text{epoch }5)

    • On MRPC: (lr 3e5,batch 16,epoch 5)(\text{lr }3e^{-5},\text{batch }16,\text{epoch }5)

    • On QNLI: (lr 3e5,batch 32,epoch 3)(\text{lr }3e^{-5},\text{batch }32,\text{epoch }3)

  • ELECTRAbase:

    • On SST-2: (lr 1e5,batch 16,epoch 5)(\text{lr }1e^{-5},\text{batch }16,\text{epoch }5)

    • On MRPC: (lr 5e5,batch 32,epoch 5)(\text{lr }5e^{-5},\text{batch }32,\text{epoch }5)

    • On QNLI: (lr 3e5,batch 32,epoch 3)(\text{lr }3e^{-5},\text{batch }32,\text{epoch }3)

These model training hyper-parameters for a specific model on one specific dataset is then fixed and reused for all the competing methods to improve backward compatibility during model upgrade.

A.2 Distillation Hyper-parameters

The knowledge distillation method from Xie et al. (2021) imposed an additional loss λKL(lold/T,lnew/T)\lambda\cdot KL(l_{old}/T,l_{new}/T) on potential regression instances. We experimented the best possible hyper-parameters from the following:

  • λ\lambda: 0.1,1.0,10.00.1,1.0,10.0

  • Temperature TT: 0.5,1.0,2.00.5,1.0,2.0

A.3 Details on Gated Fusion

We initialize the gate gθg_{\theta} to be a two-layer feed-forward network with the architecture (Dropout, Linear, LayerNorm, ReLU, Dropout, Linear, Sigmoid) and fix the hidden size to be 6464 across all our experiments.

During the training of Gated Fusion, we only train the fnewf_{new} within fGFf^{*}_{GF} for the first (N1)(N-1) epochs to ensure its competence, where NN is the total training epochs. In the last training epoch, we jointly train gθg_{\theta} and fnewf_{new} using the Gated Fusion logits lGFl^{*}_{GF} with the secondary learning rate lr2lr2. To prevent the overfitting of the gate, we also apply drop_gate where at each training step during the last epoch, there is D%D\% to only train fnewf_{new} within fGFf^{*}_{GF} and (1D)%(1-D)\% to train with lGFl^{*}_{GF}.

The hyper-parameter space of Gated Fusion is listed as follows:

  • Drop Gate (%\%): 40,50,60,8040,50,60,80

  • Temperature TT on old logits: 1.0,1.2,1.4,1.61.0,1.2,1.4,1.6

  • lr2: 5e75e^{-7}, 1e61e^{-6}, 3e63e^{-6}, 1e51e^{-5}, 3e53e^{-5}

We found that to achieve good results, the gap in logit magnitude of foldf_{old} and fnewf_{new} needs to be bridged by the temperature when upgrading from BERTbase to ELECTRAbase, with TT being 1.61.6 on SST-2, 1.61.6 on MRPC, and 1.21.2 on QNLI. On the other hand, T=1T=1 gives good results across three datasets when upgrading from BERTbase to BERTlarge. This could result from the distinct pretraining schemes between models where MLM seem to produce larger magnitude of output logits.