This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Online Prototypes and Class-Wise Hypergradients for Online Continual Learning with Pre-Trained Models

Nicolas Michel Department of Information and Communication Engineering, The University of Tokyo, Tokyo, Japan Japanese French Laboratory for Informatics, CNRS, Tokyo, Japan Maorong Wang Department of Information and Communication Engineering, The University of Tokyo, Tokyo, Japan Jiangpeng He Department of Mechanical Engineering, Massachusetts Institute of Technology, Cambridge, USA Toshihiko Yamasaki Department of Information and Communication Engineering, The University of Tokyo, Tokyo, Japan
Abstract

Continual Learning (CL) addresses the problem of learning from a data sequence where the distribution changes over time. Recently, efficient solutions leveraging Pre-Trained Models (PTM) have been widely explored in the offline CL (offCL) scenario, where the data corresponding to each incremental task is known beforehand and can be seen multiple times. However, such solutions often rely on 1) prior knowledge regarding task changes and 2) hyper-parameter search, particularly regarding the learning rate. Both assumptions remain unavailable in online CL (onCL) scenarios, where incoming data distribution is unknown and the model can observe each datum only once. Therefore, existing offCL strategies fall largely behind performance-wise in onCL, with some proving difficult or impossible to adapt to the online scenario. In this paper, we tackle both problems by leveraging Online Prototypes (OP) and Class-Wise Hypergradients (CWH). OP leverages stable output representations of PTM by updating its value on the fly to act as replay samples without requiring task boundaries or storing past data. CWH learns class-dependent gradient coefficients during training to improve over sub-optimal learning rates. We show through experiments that both introduced strategies allow for a consistent gain in accuracy when integrated with existing approaches. We will make the code fully available upon acceptance. Correspondance to Nicolas Michel, nicolas@cvm.t.u-tokyo.ac.jp

Machine Learning, Online Continual Learning, Hypergradients

1 Introduction

Continual Learning (CL) has gained significant popularity in the past decade (Kirkpatrick et al., 2017; Rao et al., 2019; Zhou et al., 2024a). The main idea relies on learning from a sequence of data rather than a fixed dataset. Consequently, the data distribution can change, and new classes can appear, which often leads to the well-known problem of Catastrophic Forgetting (French, 1999). In this paper, we focus on the Class Incremental Learning problem (Hsu et al., 2018).

Fundamentally, Continual Learning scenarios can be divided into two categories: offline Continual Learning (offCL) (Tiwari et al., 2022) and online Continual Learning (onCL) (Mai et al., 2022). The former and most popular scenario assumes that this sequence of data is clearly defined in various tasks and that the learning process on a given task is completely analogous to traditional learning. Namely, data in each task are i.i.d. and the model can be trained for several epochs on a given task before visiting the next one. On the contrary, onCL assumes that the incoming data is analogous to a stream and therefore cannot be seen more than once by the model to enforce fast adaptation. Moreover, in the objective of moving toward more realistic scenarios, various studies now consider unclear or blurry task boundaries (Moon et al., 2023; Koh et al., 2023; Bang et al., 2022) which completely disables task information usage. Such differences make methods designed for offCL seldom transferable to onCL as the vast majority relies on leveraging multiple epochs, as well as task boundary information. Recent representation-based methods such as RANPAC (McDonnell et al., 2024) or EASE (Zhou et al., 2024b) are good examples, which require clear task boundaries to compute the task-specific representations, making them incompatible with onCL. While the difficulty of not having access to task information in onCL has been addressed in various studies (Aljundi et al., 2019; Koh et al., 2023; Moon et al., 2023; Michel et al., 2024), it remains under-explored when considering Pre-Trained Models (PTM).

Another significant problem when considering the online continuous adaption of a model is the inability to estimate the optimal hyper-parameters. As the dataset of interest cannot be stored, typical hyper-parameters such as the Learning Rate (LR) should not be optimized to fully reflect the problem at hand. However, such consideration has been widely omitted in current onCL literature where LR selection strategies are often unclear. A common practice is to choose the same fixed Learning Rate (LR) and optimizer for all considered methods (Gu et al., 2022; Mai et al., 2021; Moon et al., 2023; Lin et al., 2023), typically Stochastic Gradient Descent (SGD) with LR value of 0.10.1. This design choice is extremely limited as various methods and datasets have different optimal LR. It is also well-known that a suboptimal LR has a critical impact on the final performances. Another strategy is to estimate the best hyper-parameters on a given dataset and to transfer them onto other configurations (Michel et al., 2024). While more realistic, there is no guarantee that such hyper-parameters can realistically be transferred from one dataset to another.

In this paper, we propose to improve the usage of PTM in onCL by addressing both previously introduced problems. Firstly, we solve the lack of task boundaries by simply computing Online Prototypes (OP) to recalibrate the model’s final layer towards previous classes similar to the replay strategy at the representation level. Secondly, we adapt Hypergradients (Baydin et al., 2018) to onCL by considering class-wise gradients (He, 2024); which we call Class-Wise Hypergradients (CWH). In summary, our contributions are as follows:

  • we introduce the problem of LR selection in onCL, which to the best of our knowledge remained largely unexplored;

  • we tackle onCL difficulties by leveraging Online Prototypes and introducing Class-Wise Hypergradients;

  • we demonstrate superior performances when combining such elements with state-of-the-art offCL methods, leveraging PTM in OCL across several datasets and initial LR.

2 Related Work

2.1 PTM based Continual Learning

In recent years, pre-trained models (PTMs) have been widely utilized in offCL (McDonnell et al., 2024; Lin et al., 2023; Zhou et al., 2024b; Smith et al., 2023; Wang et al., 2022b). However, their application in onCL remains largely unexplored, partly because most existing methods heavily depend on task boundaries, i.e., explicit knowledge of when the task changes. This is often assumed in pair with clear boundaries, meaning that all classes from the previous task are suddenly unavailable while all new classes encountered belong to the new task. With real-world streaming data, such a situation is equally unlikely.

2.2 Online Continual Learning

In onCL, incoming data can be seen only once, analogous to a continuous data stream (He et al., 2020). Therefore, clear boundaries are unlikely to be available and several studies suggest working in boundary-free scenarios (Buzzega et al., 2020) where task change is unknown. However, if the change is clear, it can easily be inferred. In that sense, blurry boundary setting have been proposed (Moon et al., 2023; Koh et al., 2023; Bang et al., 2022; Michel et al., 2024) in previous work. In particular, we are interested in the Si-Blurry setting (Moon et al., 2023) where not only task change is blurry, but some classes can appear or disappear during multiple tasks, which brings the experimental setup one step closer to real-world scenarios while being more challenging. While numerous studies rely on prototypes for Continual Learning (McDonnell et al., 2024; Lin et al., 2023; Zhou et al., 2024b), such representation-based methods must generally be combined with task boundary knowledge as prototypes are updated at the end of each task. In onCL, prototypes are harder to capitalize on when training a model from scratch due to the shift of representations hindering prototype computation (Caccia et al., 2022). However, when working with PTM, such a shift is drastically reduced as representations are already of high quality, making the usage of prototypes more efficient.

2.3 Hypergradients and Gradient Re-Weighting

Hypergradient (Baydin et al., 2018; Almeida et al., 1999) addresses the problem of finding the optimal learning rate in conventional training scenarios. In that sense, the authors proposed to derive a gradient descent algorithm to learn the LR. Notably, they demonstrate that computing the dot product between gradients from previous steps (θt)(θt1)\nabla\mathcal{L}(\theta_{t})\cdot\nabla\mathcal{L}(\theta_{t-1}) is sufficient to complete one step of the learning rate update rule, with tt the index of the current step, θ\theta the parameters, and \mathcal{L} the loss function. However, such techniques have been, to the best of our knowledge, developed solely for offline scenarios at a global level. In Continual Learning, gradient re-weighting strategies have been designed for replay-based CL methods. Notably, previous work proposed to re-weight the gradient at the loss level to mitigate its accumulation during training in CL context, also called gradient imbalance (Guo et al., 2023). Recently, to compensate for the class imbalance, class-wise manually defined weights in the last Fully Connected (FC) layer have been leveraged (He, 2024). Our work on Class-Wise Hypergradients lies at the cross-road between Hypergradients and Gradient Re-Weighting.

3 Learning Rate Selection in Online Continual Learning

3.1 Preliminary

Generally, the problem of CL is defined as training a model fθ()f_{\theta}(\cdot) parameterized by θ\theta on a sequence of TT tasks where each task of index k{1,,K}k\in\{1,\cdots,K\} is defined by its corresponding dataset 𝒟k\mathcal{D}_{k}, each potentially being drawn from a different distribution. In the case Class Incremental Learning (Hsu et al., 2018), we have 𝒟k=(𝒳k,𝒴k)\mathcal{D}_{k}=(\mathcal{X}_{k},\mathcal{Y}_{k}), the data-label pairs. In offCL, the model is trained sequentially on each task while other task data is unavailable. For onCL, while the model is similarly trained sequentially on each task, only the data of the current batch is available and can be seen only once during training. The final objective in offCL and onCL is the same: to maximize performance across all tasks. Equivalently, to minimize the average loss {\mathcal{L}} across all tasks:

minθ1Tk=1T(θ,𝒟k).\min_{\theta}\frac{1}{T}\sum_{k=1}^{T}\mathcal{L}(\theta,\mathcal{D}_{k}). (1)

3.2 Finding the Optimal Learning Rate

Traditional Learning.

Finding the optimal LR is a common challenge inherent to training any deep neural network with gradient-descent-based optimization techniques. Mainstream methods heavily rely on LR schedulers (He et al., 2016; Vaswani, 2017), which would typically decrease the LR value over time after each epoch. Of course, the starting value as well as the speed of the learning rate decrease must still be found. To this end, grid search remains a popular and powerful technique in traditional learning scenarios.

LR in offCL.

Similarly, in offCL, state-of-the-art methods have adopted LR schedulers (Smith et al., 2023; Wang et al., 2022b; Roy et al., 2024), however their strategy regarding finding hyperparameters is often unclear. While grid search is feasible in offCL at the task level, clear limitations can be identified. Firstly, finding the best initial LR at a given task kk on dataset 𝒟k\mathcal{D}_{k} does not give any guarantee regarding the LR to use on subsequent datasets 𝒟k+1,,𝒟T\mathcal{D}_{k+1},\cdots,\mathcal{D}_{T}. Secondly, a suboptimal learning rate with regard to 𝒟k\mathcal{D}_{k} can lead to overall higher performances when evaluating on {𝒟0,,𝒟k1}\{\mathcal{D}_{0},\cdots,\mathcal{D}_{k-1}\}. Indeed, as discussed in the work of Mirzadeh et al. (Mirzadeh et al., 2020), large learning rates can disturb weights that are important for previous tasks. The loss minimized at specific timestep conflicts with the overall objective since the distribution of incoming data from 𝒟k\mathcal{D}_{k} and k𝒟k\bigcup_{k}\mathcal{D}_{k} differ.

LR in onCL.

Determining the optimal learning rate in onCL is even more challenging. In addition to the offCL challenges mentioned earlier, even grid search at the task level is unavailable. The online scenario constrains the model to see the incoming data only once, usually in small batches, making any hyper-parameter search completely impossible. A long-lasting strategy to train in this context is to use the same pre-determined fixed LR and optimizer for all considered methods (Gu et al., 2022; Mai et al., 2021; Moon et al., 2023; Lin et al., 2023), typically Stochastic Gradient Descent (SGD) with an LR value of 0.10.1. While the choice of SGD has been found to lead to better generalization in some cases (Keskar & Socher, 2017), the reasoning behind this choice remains unclear in the literature and presents significant limitations. Firstly, there is no justification for different methods, with different losses, to benefit equally from using the same learning and optimizer. Secondly, as discussed in Section 5.3, it is apparent that some methods are more sensitive to LR change than others. To overcome this, some studies look for the best hyper-parameters offline for all considered methods before transferring such parameters to other setups (Michel et al., 2024). However, this remains insufficient as the optimal learning rate most likely varies from one dataset to the other. Despite the well-known critical impact of a suboptimal LR on final performance, the problem of selecting the best LR remains largely overlooked by the Online Continual Learning community.

3.3 LR and Learning Behavior in onCL

Impact on Stability-Plasticity.

It is clear that selecting an appropriate learning rate is essential for optimal performance. In standard scenarios, the impact of its choice on loss minimization and convergence speed has been extensively studied (Ruder, 2016). For offCL, previous studies have considered to impact of the LR on forgetting (Mirzadeh et al., 2020). Notably, a higher LR would increase forgetting, and vice-versa. Intuitively, the learning rate gives a direct control on the plasticity-stability tradeoff (Wang et al., 2024). To confirm such behavior in onCL, we experiment with larger and smaller LR values. As it can be seen in Figure 1, when trained with a higher learning rate (5×1025\times 10^{-2}), the model tends to obtain higher performances on the latest tasks while exhibiting especially low performances on earlier tasks. When trained with a lower LR (5×1055\times 10^{-5}), the model tends to achieve better performance on earlier tasks compared to training with a higher LR. In other words, a high LR value induces more plasticity and less stability, and vice-versa.

Refer to caption
Figure 1: Task-wise Accuracy (%) of DualPrompt at the end of training on CIFAR100, split in 10 tasks, for LR values in {5×105,5×102}\{5\times 10^{-5},5\times 10^{-2}\}

FC Layer Behavior.

For PTM in offCL, previous work considers different LR values for different layers (McDonnell et al., 2024). Notably, leveraging a higher LR in the last FC layer seems beneficial to the training procedure. In onCL, we observe similar behavior. To do so, we experiment with CODA (Smith et al., 2023) where we multiply the gradient values with regard to the final layer by a coefficient κ{1,,1000}\kappa\in\{1,\cdots,1000\}, with κ=1\kappa=1 falling back to the regular CODA training in onCL. Such results are presented in Table 1 on various onCL datasets. It can be seen that increasing the LR of the final FC layer can often lead to an improvement in terms of average performances, even though in cases of lower LR such an effect is barely noticeable. While this naive approach can lead to improvements in certain scenarios, it induces more hyper-parameter tuning and considers the same LR strategy for each class.

Table 1: Average Performances (%) of CODA on all datasets for various LR values with and with multiplying the final FC layer learning rate by a coefficient κ\kappa.
Learning Rate 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3}
CIFAR100
CODA (κ=1)\kappa=1) 73.36±1.88 81.69±1.48 64.98±3.47
CODA (κ=10\kappa=10) 73.40±1.88 81.74±2.16 66.90±0.47
CODA (κ=100\kappa=100) 73.41±1.84 81.43±1.94 64.10±3.41
CODA (κ=1000\kappa=1000) 73.41±1.88 81.75±0.87 67.93±1.63
ImageNet-R
CODA (κ=1)\kappa=1) 24.62±1.67 64.03±1.81 33.18±3.59
CODA (κ=10\kappa=10) 24.62±1.66 64.93±0.85 31.89±4.07
CODA (κ=100\kappa=100) 24.62±1.66 65.25±0.21 36.85±8.25
CODA (κ=1000\kappa=1000) 24.62±1.66 64.87±0.62 35.32±5.44
CUB
CODA (κ=1)\kappa=1) 22.06±0.87 65.28±3.07 40.25±12.5
CODA (κ=10\kappa=10) 22.03±0.9 65.70±2.48 51.15±0.94
CODA (κ=100\kappa=100) 22.03±0.9 65.06±2.07 48.26±1.7
CODA (κ=1000\kappa=1000) 22.05±0.89 65.70±2.81 45.32±6.52

4 Proposed Method

4.1 Motivations

As discussed in Section 3, LR is critical in onCL as it not only has a direct impact on the plasticity-stability trade-off, it is near impossible to use the optimal value. While previous studies highlight the need for different LR at the layer level, the case of different learning rates at the class level is still in its infancy. Following the analysis described in Section 3.3, we make the hypothesis that a class-wise LR strategy focusing on the last FC layer is crucial for onCL. The intuition is that different classes require different learning rates so that the model can adapt its stability-plasticity tradeoff not only over time, but also over classes. In that sense, we reckon hypergradients (Baydin et al., 2018) to be adequate for improving over non-optimal initial LR values. However, hypergradients are not designed for the onCL scenario and a naive implementation leads to a severe performance drop. Indeed, hypergradient computation is identical regardless of the network layer or the classes. Inspired by the work of He et al. (He, 2024), we propose learning how to rescale the gradient coefficient of the last FC layer, class-wise, using hypergradient learning theory.

Additionally, to adapt PTM-based methods to the onCL scenario further, we leverage the stability of the output representation of PTM by computing Online Prototypes.

4.2 Class-Wise Hypergradients

Let us consider a model fθf_{\theta} parameterized by θ\theta such that for an input xdx\in\mathbb{R}^{d}, with dd the dimension of the input space, we have fθ(x)=hw(x)TWf_{\theta}(x)=h_{w}(x)^{T}\cdot W, with Wl,cW\in\mathbb{R}^{l,c}, cc the number of classes, ll the dimension of the output of hwh_{w} and θ={w,W}\theta=\{w,W\}. In this context, hwh_{w} would typically be a PTM and WW the weight of the final FC layer (including the bias). Looking at the final FC layer WW, for a learning rate η\eta, we can write the gradient-based weight update:

Wt+1=Wtη(Wt),W_{t+1}=W_{t}-\eta\nabla\mathcal{L}(W_{t}), (2)

with tt the iteration index. Here, we omit the input data for simplicity. Then, for any class index j{1,c}j\in\{1,c\}, we can write the update rule of the weights corresponding to jj:

Wt+1j=Wtjη(Wtj),W_{t+1}^{j}=W_{t}^{j}-\eta\nabla\mathcal{L}(W_{t}^{j}), (3)

with WjW^{j} the jthj^{th} column of WW. Building upon previous studies (He, 2024), we introduce step-dependent class-wise weighting coefficients, leading to the following update rule:

Wt+1j=Wtjαt+1jη(Wtj),W_{t+1}^{j}=W_{t}^{j}-\alpha_{t+1}^{j}\eta\nabla\mathcal{L}(W_{t}^{j}), (4)

with αtj+\alpha_{t}^{j}\in\mathbb{R}^{+\star}, class dependent gradient weighting coefficient at index tt. While those coefficients were traditionally introduced to compensate for class interference, and computed with hand-crafted rules, we propose to learn them through an online adaptive rule based on hypergradients theory. In particular, we want to construct a higher level update for {αtj}j\{\alpha_{t}^{j}\}_{j} such that, in the case:

αt+1j=αtjβ(Wtj)αj,\alpha_{t+1}^{j}=\alpha_{t}^{j}-\beta\frac{\partial\mathcal{L}(W_{t}^{j})}{\partial\alpha^{j}}, (5)

with β+\beta\in\mathbb{R}^{+} the hypergradient learning rate. To compute the partial derivative we apply the chain rule and make use of the fact that Wtj=Wt1jαjη(Wt1j)W^{j}_{t}=W^{j}_{t-1}-\alpha^{j}\eta\nabla\mathcal{L}(W_{t-1}^{j}), such that:

(Wtj)αj\displaystyle\frac{\partial\mathcal{L}(W_{t}^{j})}{\partial\alpha^{j}} =(Wtj)Wtjαj\displaystyle=\nabla\mathcal{L}(W_{t}^{j})\cdot\frac{\partial W_{t}^{j}}{\partial\alpha^{j}} (6)
=η(Wtj)(Wt1j).\displaystyle=-\eta\nabla\mathcal{L}(W_{t}^{j})\cdot\nabla\mathcal{L}(W_{t-1}^{j}). (7)

The resulting Class-Wise Hypergradient update becomes:

αt+1j=αtj+γ(Wtj)(Wt1j).\alpha_{t+1}^{j}=\alpha_{t}^{j}+\gamma\cdot\nabla\mathcal{L}(W_{t}^{j})\cdot\nabla\mathcal{L}(W_{t-1}^{j}). (8)

With γ=βη\gamma=\beta\eta. Naturally, this introduces an undesired extra-hyperparameter. We discuss this limitation in Section 6. For clarity, the relation presented in Eq. (8) relies on an SGD update. In practice, we favor a momentum-based update, notably Adam (Kingma & Ba, 2014). We follow the guidelines of Baydin et al. (Baydin et al., 2018) regarding its implementation and give more detail in Appendix A.

4.3 Online Prototypes

In order to reduce forgetting in the last layer, we compute Online Prototypes (OP) 𝒫={pk11,pk22,,pkcc}\mathcal{P}=\{p_{k_{1}}^{1},p_{k_{2}}^{2},\cdots,p_{k_{c}}^{c}\} of each class during training. For a given class jj, the class prototype pkjjp_{k_{j}}^{j} computed over kjk_{j} samples is updated when encountering a new sample xkj+1jx_{k_{j}+1}^{j}. For simplicity, we omit the jj index in kjk_{j} going forward. Therefore, the prototype update rule is:

pk+1j=kpkj+hw(xk+1j)k+1,p^{j}_{k+1}=\frac{k\cdot p_{k}^{j}+h_{w}(x_{k+1}^{j})}{k+1}, (9)

with xk+1jx_{k+1}^{j} the k+1th{k+1}^{th} encountered sample of class jj. For all classes, prototypes are initialized such that p0j=0p_{0}^{j}=0. Prototypes are then used to recalibrate the final FC layer, analogous to replaying the average of past data representation during training. In that sense, we define the prototype-based loss term as:

OP=-1cj𝒞oldlog((pj)TWj),\mathcal{L}_{OP}=\frac{\text{-}1}{c}\sum_{j\in\mathcal{C}_{old}}\log\left((p^{j})^{T}\cdot W^{j}\right), (10)

with 𝒞old={j{1,c}|pkjj0}\mathcal{C}_{old}=\{j\in\{1,c\}\ |\ p^{j}_{k_{j}}\neq 0\}. OP\mathcal{L}_{OP} is the cross-entropy with regard to prototypes of encountered classes. OP acts as simple and low-budget memory data.

4.4 Overall Training Procedure

The approach relies on two components that are orthogonal to most existing methods, as long as the model optimizes a final FC layer for classification. In that sense, we propose to integrate it into various state-of-the-art baselines relying on PTM, notably prompt-based approaches. To do so, if we consider base\mathcal{L}_{base} to be the loss of the baseline method, to which we attach both components OP and CWH, we have the overall loss to minimize:

=base+OP.\mathcal{L}=\mathcal{L}_{base}+\mathcal{L}_{OP}. (11)

A Pytorch-like (Paszke et al., 2019) pseudo code is given in Algorithm 1. Additionally, we leverage batch-wise masking to consider the logits of classes that are only present in the current batch. More details can be found in the Appendix A. Similarly, note that the bias has been omitted for simplicity.

gamma = 1e-3
grad_weight = torch.ones(n_classes)
prev_grad = None
for x, y in dataloader:
# Baseline loss
h, y_hat = network(x) # features and logits
loss_baseline = criterion_baseline(y_hat, y)
# FC recalibration
proto, labels = get_prototypes()
logits = network.fc(proto)
loss_op = cross_entropy(logits, labels) # Eq. 10
\parloss = loss_baseline + loss_op # Eq. 11
\paroptim.zero_grad()
loss.backward()
\par# Class-Wise Hypergradient update
curr_grad = network.fc.weight.grad
if prev_grad is not None:
grad_weight += gamma * (curr_grad @ prev_grad.T).diag() #Eq. 8
for i in range(n_classes):
network.fc.weight.grad[i, :] *= grad_weight[i]
prev_grad = curr_grad
\paroptim.step()
update_proto(h, y) # Eq. 9
Algorithm 1 PyTorch-like pseudo-code of our method integration with other baselines

5 Experiments

In the following sections experiment with combining our proposed approach with several flagship CL approaches.

Table 2: Average Performances (%) of all considered baselines with and without our proposed OP and CWH, in the clear setting. Results are displayed for different datasets and LR values. We also report the average across LR. We used 3 runs and reported the average and standard deviations. Values in blue are the best for each column. Bold are the best on average. Best viewed in color.
Dataset CIFAR100 Imagenet-R CUB
Learning Rate 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg. 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg. 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg.
L2P 67.86±2.12 78.18±1.29 68.54±0.98 71.53±1.46 25.93±3.87 55.56±1.81 50.18±1.36 43.89±2.35 20.15±2.69 55.21±2.22 51.27±2.09 42.21±2.33
     \hookrightarrow + OP 80.74±1.08 85.73±0.06 79.11±0.87 81.86±0.67 47.96±1.2 68.68±1.13 60.57±0.96 59.07±1.1 48.20±0.84 83.51±1.49 73.29±1.2 68.33±1.18
        \hookrightarrow + CWH 81.73±0.93 85.52±0.1 78.16±1.43 81.80±0.82 53.18±1.68 68.89±1.11 58.38±2.01 60.15±1.6 58.84±0.9 83.98±1.3 72.78±2.1 71.87±1.43
DualPrompt 62.21±1.21 77.36±0.46 65.10±0.92 68.22±0.86 25.34±1.32 59.11±0.54 53.15±3.01 45.87±1.62 21.77±3.13 65.34±1.22 53.59±2.17 46.90±2.17
     \hookrightarrow + OP 78.16±1.31 85.35±0.76 68.28±1.95 77.26±1.34 49.10±1.36 67.68±0.37 57.40±2.1 58.06±1.28 55.06±1.34 86.13±0.84 69.72±1.69 70.30±1.29
        \hookrightarrow + CWH 79.78±1.29 85.08±0.68 67.27±1.77 77.38±1.25 54.59±1.07 67.55±0.51 55.03±4.77 59.06±2.12 66.71±0.95 86.67±0.91 69.10±3.73 74.16±1.86
CODA 73.36±1.88 81.69±1.48 64.98±3.47 73.34±2.28 24.62±1.67 64.03±1.81 33.18±3.59 40.61±2.36 22.06±0.87 65.28±3.07 40.25±12.5 42.53±5.48
     \hookrightarrow + OP 85.74±0.7 86.16±0.64 72.42±0.8 81.44±0.71 49.39±0.87 72.30±1.7 38.72±3.57 53.47±2.05 55.35±1.66 81.21±2.33 59.59±8.63 65.38±4.21
        \hookrightarrow + CWH 86.92±0.66 85.45±1.37 70.11±0.49 80.83±0.84 55.63±0.97 70.12±2.17 41.32±5.62 55.69±2.92 66.74±1.79 81.78±2.28 60.26±3.09 69.59±2.39
ConvPrompt 78.61±1.25 85.76±0.43 74.40±0.69 79.59±0.79 30.98±0.97 66.66±1.33 17.45±27.19 38.36±9.83 31.16±1.04 70.78±2.65 34.13±16.8 45.36±6.83
     \hookrightarrow + OP 88.70±0.77 90.41±0.54 78.59±6.67 85.90±2.66 53.82±1.41 74.79±1.13 21.60±26.5 50.07±9.68 57.42±3.4 86.66±1.35 42.19±29.36 62.09±11.37
        \hookrightarrow + CWH 88.86±0.63 90.16±0.21 80.94±5.31 86.65±2.05 53.17±1.0 72.94±1.31 34.90±28.87 53.67±10.39 65.02±3.77 84.56±0.75 65.42±19.02 71.67±7.85
MVP 49.64±0.92 57.63±2.13 43.40±1.02 50.22±1.36 28.75±0.42 38.25±0.54 28.87±0.79 31.96±0.58 28.39±0.81 44.10±1.63 29.96±0.81 34.15±1.08
     \hookrightarrow + OP 82.63±0.91 85.61±0.12 56.16±1.3 74.80±0.78 58.04±1.05 62.08±0.29 47.50±1.9 55.87±1.08 72.06±2.05 85.86±1.63 59.64±4.41 72.52±2.7
        \hookrightarrow + CWH 83.64±0.59 85.44±0.11 61.60±1.27 76.89±0.66 60.39±0.74 62.25±0.35 50.09±3.7 57.58±1.6 77.15±2.0 85.21±1.44 70.49±3.35 77.62±2.26
Table 3: Average Performances (%) of all considered baselines with and without our proposed OP and CWH, in the Si-Blurry setting. Results are displayed for different datasets and LR values. We also report the average across LR. We used 3 runs and reported the average and standard deviations. Values in blue are the best for each column. Bold are the best on average. Best viewed in color.
Dataset CIFAR100 Imagenet-R CUB
Learning Rate 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg. 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg. 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg.
L2P 47.37±6.28 71.45±6.51 67.35±4.82 62.06±5.87 13.71±2.27 44.02±4.47 44.1±0.84 33.94±2.53 7.51±2.91 46.57±1.18 47.25±2.27 33.78±2.12
     \hookrightarrow + OP 69.68±5.59 82.05±4.96 79.25±2.35 76.99±4.3 32.99±2.23 61.34±1.93 55.6±2.49 49.98±2.22 24.46±6.64 76.53±1.7 69.52±2.46 56.84±3.6
        \hookrightarrow + CWH 73.01±5.23 82.39±3.66 78.17±2.35 77.86±3.75 39.22±2.03 60.99±1.84 54.41±1.8 51.54±1.89 36.45±6.31 77.0±1.97 67.7±2.13 60.38±3.47
DualPrompt 46.07±3.25 72.03±5.63 65.72±5.12 61.27±4.67 15.29±0.68 50.13±1.32 45.84±2.81 37.09±1.6 12.46±1.5 57.03±0.4 51.47±1.25 40.32±1.05
     \hookrightarrow + OP 66.35±9.42 82.15±2.72 72.3±2.01 73.6±4.72 36.79±1.05 59.78±0.95 51.74±2.81 49.44±1.6 37.91±2.38 79.64±1.99 67.89±1.59 61.81±1.99
        \hookrightarrow + CWH 72.41±3.65 82.06±2.32 71.91±1.42 75.46±2.46 42.46±0.91 59.49±0.81 50.56±2.17 50.84±1.3 50.26±2.25 80.48±1.83 66.59±2.07 65.78±2.05
CODA 54.38±3.63 77.23±7.11 64.04±5.82 65.22±5.52 15.89±2.47 56.38±0.51 28.47±2.74 33.58±1.91 12.52±1.65 57.6±1.34 45.79±1.86 38.64±1.62
     \hookrightarrow + OP 75.02±2.02 84.0±4.3 73.12±1.29 77.38±2.54 37.85±1.92 64.55±0.39 31.84±14.84 44.75±5.72 36.83±1.88 76.25±0.29 60.6±4.42 57.89±2.2
        \hookrightarrow + CWH 77.71±2.94 84.39±3.56 74.61±1.69 78.9±2.73 43.45±1.6 63.53±1.72 34.93±1.11 47.3±1.48 49.24±1.16 77.53±1.32 56.81±2.2 61.19±1.56
ConvPrompt 42.76±1.76 48.18±1.43 26.54±22.0 39.16±8.4 15.02±0.43 33.17±1.82 6.2±9.66 18.13±3.97 13.93±1.78 32.0±0.71 21.33±6.09 22.42±2.86
     \hookrightarrow + OP 49.96±2.01 51.65±2.39 47.29±2.43 49.63±2.28 28.79±2.4 38.75±1.62 17.32±14.38 28.29±6.13 30.49±2.79 43.42±2.45 11.18±10.9 28.36±5.38
        \hookrightarrow + CWH 49.72±2.03 51.95±2.13 47.47±2.32 49.71±2.16 28.0±2.7 38.35±2.09 16.92±13.97 27.76±6.25 33.79±3.34 42.3±1.46 36.74±3.2 37.61±2.67
MVP 44.32±2.27 60.96±5.42 52.29±4.65 52.52±4.11 23.65±0.56 35.5±0.16 32.54±0.91 30.56±0.54 23.87±2.24 42.85±0.37 34.64±1.72 33.79±1.44
     \hookrightarrow + OP 77.16±4.91 82.52±2.78 62.56±2.99 74.08±3.56 51.21±0.65 56.94±1.07 46.49±2.16 51.55±1.29 65.55±2.24 80.93±0.97 61.42±0.21 69.3±1.14
        \hookrightarrow + CWH 78.49±4.89 82.46±2.58 64.12±1.73 75.02±3.07 53.75±0.76 57.02±1.03 45.5±2.21 52.09±1.33 71.81±1.48 81.46±1.16 63.27±0.6 72.18±1.08

5.1 Metrics

Average Performances (AP).

We follow previous work and define the Average Performance (AP) as the average of the accuracies computed after each task during training (Zhou et al., 2024a). Formally, when training on {𝒟1,,𝒟T}\{\mathcal{D}_{1},\cdots,\mathcal{D}_{T}\}, we define 𝒜k\mathcal{A}_{k}, the Average Accuracy (AA) as:

𝒜k=1kl=1kal,k\mathcal{A}_{k}=\frac{1}{k}\sum_{l=1}^{k}a_{l,k} (12)

with al,ka_{l,k} the accuracy on task ll after training on 𝒟k\mathcal{D}_{k}. Building on this, we define the Average Performance (AP) as:

𝒫=1Tk=1T𝒜k\mathcal{P}=\frac{1}{T}\sum_{k=1}^{T}\mathcal{A}_{k} (13)

Performance Across LR.

To show the improvement in the case of unknown optimal LR, we propose to experiment with various LR values and report individual and averaged performances across these values. Specifically, we experiment for LR values in {5×105,5×104,5×103}\{5\times 10^{-5},5\times 10^{-4},5\times 10^{-3}\}. The main motivation is that we reckon that the optimal LR is likely to fall into that range, therefore we wanna take into account the effect of experimenting with a learning rate that is either above or below the optimal value. Such a metric should emphasize the validity of the approach when the optimal LR is unknown and result in a fairer comparison than using the same LR blindly for every approach.

5.2 Experimental Setting

Baselines and Datasets.

In order to demonstrate the efficiency of our approach as presented in Algorithm 1, we integrate it with several state-of-the-art methods in offCL. Notably, L2P (Wang et al., 2022b), DualPrompt (Wang et al., 2022a), CODA (Smith et al., 2023), ConvPrompt (Roy et al., 2024). These methods are not naturally suited for the online case, so they had to be adapted. More details on the adaptation of such methods are in Appendix D. Additionally, we include one state-of-the-art onCL method that leverages PTM, MVP (Moon et al., 2023). We evaluate our method on CUB (Wah et al., 2011), ImageNet-R (Hendrycks et al., 2021) and CIFAR100 (Krizhevsky, 2012). More details in Appendix B.2.

Clear Boundaries.

We experiment in clear boundaries settings, for continuity with previous work, despite its lack of realism for onCL. In that sense, we consider an initial count of 1010 classes for the first task, with an increment of 1010 classes per task. This results in 1010 tasks with 1010 classes per task for CIFAR100, as well as 2020 tasks with 1010 classes per task for CUB and ImageNet-R.

Blurry Boundaries.

To evaluate our method in more realistic scenarios, we reckon the Si-Blurry (Moon et al., 2023) setting to be the most relevant to our study case. Specifically, we use their implementation of Stochastic incremental Blurry boundaries (Si-Blurry). We use the same number of tasks as for the clear setting. In this case, some classes can appear or disappear during training and the transitions are not necessarily clear. More details on this setting can be found in Appendix E. To adequately adapt the proposed methods to the online problem, we use batch-wise masking. Specifically, we consider only the logits of the class that are observed in the current batch, for every method.

Implementation Details.

Every method is evaluated in the onCL context. Namely, only one pass over the data is allowed. The batch size is fixed at 1010 to simulate small data increments with a low storage budget in the context of fast adaptation. The PTM used is a ViT-base, pre-trained on ImageNet 21k. Every experiment was conducted over 3 runs and the average and standard deviation are displayed. Each run was conducted with a different seed, which equally impacted the task generation process. For all experiments, we use γ=1×103\gamma=1\times 10^{-3}. More details on γ\gamma selection can be found in Section 6.2

5.3 Experimental Results

We combined our method with four offline approaches and one online state-of-the-art approach, all using PTM. As mentioned in Section 4.4, our method must be applied on the classification head, therefore prompt-based approaches are especially suited as they all leverage a final FC layer on top of the PTM representation for classification.

Average Performances.

We experiment in both clear and blurry settings and present the results in terms of Average Performance in Table 2 and Table 3. On top of datasets and boundary scenarios, we propose a novel evaluation procedure specific to the problem at hand. Namely, we present results for LR values in {5×105,5×104,5×103}\{5\times 10^{-5},5\times 10^{-4},5\times 10^{-3}\}, as well as the AP across all these values. The objective is to observe how would the method perform on average when the optimal learning rate is unknown and might be far from optimal, by being either too high or too low. Therefore, in all cases, combining both proposed components leads to a significant improvement in AP over the baselines when looking at the average across learning rates, with up to 30%30\% improvement on CUB. Such improvements are observed in both blurry and clear scenarios, confirming the ability of our approach to perform in realistic and traditional contexts. Individual contributions of OP and CWH are detailed below.

Blurry Boundaries.

Even though performance improvement can be observed as well in the blurry scenario, all methods suffer from a significant drop when transitioning from one scenario to the other. Such behavior highlights the importance of focusing on more realistic setups in future research. In that sense, our method remains completely applicable regardless of the presence of boundaries.

Ablation Study.

To make apparent the contribution of each component of our method, we included the performances of the original baselines, followed by the performance of such baselines combined with OP (+ OP) and eventually the performance of the same baseline combined with prototypes and CWH (+ CWH). Such results are included in Tables 2 and 3. While it is clear that the usage of prototypes is largely beneficial, in some situations the addition of CWH can lead to a slight drop in performances. However, in those rare scenarios, the performance drop remains minimal with the maximum drop value being 3.79%3.79\% in the case of CODA on CUB with an LR of 5×1035\times 10^{-3}. In other cases, performance loss is around 1%1\%. Despite this limitation, it is important to note that the gain of including CWH is especially important for longer task sequences such as CUB, which is a more realistic scenario. Moreover, when the initial LR value is particularly low (5×1055\times 10^{-5}), the gain of including CWH is generally more important. When fine-tuning PTM, it is common to start with low LR values. This property further confirms the ability of our approach to perform in realistic contexts.

Table 4: AP for L2P on CIFAR100, clear setup, with and without OP, with and without freezing weights, and without CWH.
Initial LR w/o OP w/ OP
Not Frozen 5×1045\times 10^{-4} 8.97 8.18
5×1055\times 10^{-5} 50.64 51.14
Frozen 5×1045\times 10^{-4} 77.10 85.73
5×1055\times 10^{-5} 67.86 80.74

6 Discussions

6.1 Stability of OP

Despite its apparent simplicity, OP gives the largest performance gain. This is partially due to the fact that only the FC layer and the input prompts are trained, so the output prototypes are stable over time. In Table 4 we show that unfreezing intermediate weight not only drastically decreases overall performances but similarly negates the effect of OP significantly. Additionally, in Figure 2, we show the average Euclidean distance of Online Prototypes between two training step iterations. As we can see, the learned prototypes tend to stabilize rapidly over time, even more so when the intermediate weights remain frozen. This stability of the prototype indicates that even when computed online, they can be used as a reasonable proxy for class average representation.

Refer to caption
Figure 2: Prototype shift of L2P with LR of 5×1055\times 10^{-5} with and without freezing the network weights.

6.2 Selecting γ\gamma

The main drawback of leveraging CWH is the addition of an extra hyper-parameter γ\gamma, whereas our objective is to reduce the dependency on hyper-parameters for onCL. However, we argue that selecting γ\gamma remains particularly simple. To demonstrate this, we experiment with γ{0,1×105,1×103,1}\gamma\in\{0,1\times 10^{-5},1\times 10^{-3},1\}, for an initial LR η{5×105,5×103}\eta\in\{5\times 10^{-5},5\times 10^{-3}\}. In that case, γ=0\gamma=0 is equivalent to disabling CWH. Such results are presented in Table 5. When the initial LR is lower, including CWH leads to an improvement in performance, especially when γ\gamma is large. Experimentally, this can be seen as a consequence of {αj}j\{\alpha^{j}\}_{j} values increasing during training in all cases, as discussed in Section 6.3. Therefore, it is natural that lower initial LR values would benefit more from such a strategy. When the initial LR is large, as discussed in Section 5.3, a marginal drop in performance can be observed in some cases. However, the final performances remain stable for all values of γ\gamma, minimizing the need for hyperparameter search. Additionally, the lower the value of γ\gamma, the closer it is to the original method.

In conclusion, for higher LR we recommend lower values of γ\gamma, as their effect on the training is reduced. For lower LR, any value of γ\gamma leads to an improvement, however, higher values are recommended for optimal accuracy gain. Default values such as 0.0010.001 should lead to an increase in performances in most cases or a slight decrease in worst cases. Overall, selecting γ\gamma remains a simpler and more realistic task than doing a grid search in onCL.

Table 5: Impact of γ\gamma for initial LR values η{5×103,5×105}\eta\in\{5\times 10^{-3},5\times 10^{-5}\}. All methods are trained in the Si-Blurry setting, with OP.
γ\gamma 0 1×1051\times 10^{-5} 1×1031\times 10^{-3} 11
CIFAR100 η=5e-5\eta=5e\text{-}5 CODA 75.02±2.02 76.04±2.29 77.71±2.94 77.75±2.99
DualPrompt 66.35±9.42 70.57±3.7 72.41±3.65 72.45±3.65
L2P 69.68±5.59 71.11±5.56 73.01±5.23 73.10±5.19
η=5e-3\eta=5e\text{-}3 CODA 73.12±1.29 72.88±1.72 74.61±1.69 75.77±2.91
DualPrompt 72.30±2.01 71.79±2.68 71.91±1.42 73.42±1.08
L2P 79.25±2.35 78.85±2.09 78.17±2.35 78.31±1.13
ImageNet-R η=5e-5\eta=5e\text{-}5 CODA 37.85±1.92 40.09±1.8 43.45±1.6 43.51±1.58
DualPrompt 36.79±1.05 39.06±1.07 42.46±0.91 42.54±0.89
L2P 32.99±2.23 35.65±2.15 39.22±2.03 39.30±2.0
η=5e-3\eta=5e\text{-}3 CODA 31.84±14.84 38.46±1.9 34.93±1.11 33.47±2.7
DualPrompt 51.74±2.81 50.86±2.3 50.56±2.17 50.05±2.2
L2P 55.60±2.49 55.25±1.55 54.41±1.8 54.80±1.12
CUB η=5e-5\eta=5e\text{-}5 CODA 36.83±1.88 40.28±1.95 49.24±1.16 49.40±1.07
DualPrompt 37.91±2.38 41.26±2.59 50.26±2.25 50.46±2.16
L2P 24.46±6.64 27.88±6.59 36.45±6.31 36.63±6.38
η=5e-3\eta=5e\text{-}3 CODA 60.60±4.42 56.58±4.14 56.81±2.2 58.42±6.1
DualPrompt 67.89±1.59 66.8±1.92 66.59±2.07 67.47±2.35
L2P 69.52±2.46 68.86±2.25 67.70±2.13 67.55±2.75

6.3 Values of learned αj\alpha^{j}

Trend and Distribution Across Classes.

The class-wise coefficients αj\alpha^{j} are learned in each training step as described in Eq. (8). We report coefficients’ evolution in Figure 3. Firstly, learned coefficient values are superior to 11 and increase during training. Such a trend is aligned with findings from previous studies suggesting adopting a higher LR value for the final FC layer (McDonnell et al., 2024). Secondly, it can be seen that coefficient values tend to be larger for later tasks than for earlier tasks. Intuitively, following the analysis from Section 3, this corresponds to giving more plasticity to newly encountered classes than older classes.

Refer to caption
Figure 3: Average of αj\alpha^{j} values per tasks when training CODA + CWH on CIFAR-100 with clear boundaries, γ=0.001\gamma=0.001, and LR of 0.0050.005. 0-10 corresponds to the average of αj,j{1,10}\alpha^{j},\ j\in\{1,10\} and so on. Here, classes are given in order, so 0-10 is the average for all classes corresponding to the first task.

Relation with Learning Rate.

We investigate the behavior of the learned coefficient with regard to the initial LR. As we can see in Figure 4, the smaller the initial LR, the larger the values of learned coefficients. Intuitively, a smaller initial LR requires a larger compensation in the FC layer to reach a similar learning regime.

Refer to caption
Figure 4: Average of αj\alpha^{j} values on the first task when training CODA+CWH on CIFAR-100 with clear boundaries, γ=0.001\gamma=0.001 for various initial learning rates.

7 Conclusion

In this paper, we studied the problem of leveraging Pre-Trained Models in the context of Online Continual Learning. In that sense, we focused on two central problems. Namely, the unavailability of task boundaries and the unfeasibility of hyper-parameter search. We tackled the former by leveraging Online Prototypes, a simple yet powerful strategy that takes advantage of the stable representation output of PTM to use class prototypes as replay samples. For the latter, we introduced Class-Wise Hypergradients which can help mitigate the drop in performances due to inadequate LR value. To reflect the efficiency of our approach, we evaluate every model with various initial LR values and show that on average both OP and CWH can be beneficial when combined with baseline models. Nonetheless, the problem of using optimal hyper-parameters in an online context remains unsolved and we hope that this work can shed light on its importance and pave the way to additional research in this direction.

Impact Statement

This paper presents work whose goal is to advance the field of Machine Learning. There are many potential societal consequences of our work, none of which we feel must be specifically highlighted here.

References

  • Aljundi et al. (2019) Aljundi, R., Kelchtermans, K., and Tuytelaars, T. Task-free continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  11254–11263, 2019.
  • Almeida et al. (1999) Almeida, L. B., Langlois, T., Amaral, J. D., and Plakhov, A. Parameter adaptation in stochastic optimization. In On-line learning in neural networks, pp.  111–134, 1999.
  • Bang et al. (2022) Bang, J., Koh, H., Park, S., Song, H., Ha, J.-W., and Choi, J. Online continual learning on a contaminated data stream with blurry task boundaries. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  9275–9284, 2022.
  • Baydin et al. (2018) Baydin, A. G., Cornish, R., Rubio, D. M., Schmidt, M., and Wood, F. Online learning rate adaptation with hypergradient descent. In Sixth International Conference on Learning Representations, 2018.
  • Buzzega et al. (2020) Buzzega, P., Boschini, M., Porrello, A., Abati, D., and Calderara, S. Dark experience for general continual learning: a strong, simple baseline. In Advances in Neural Information Processing Systems, volume 33, pp.  15920–15930, 2020.
  • Caccia et al. (2022) Caccia, L., Aljundi, R., Asadi, N., Tuytelaars, T., Pineau, J., and Belilovsky, E. New insights on reducing abrupt representation change in online continual learning. In International Conference on Learning Representations, 2022.
  • Dosovitskiy et al. (2021) Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., and Houlsby, N. An image is worth 16x16 words: Transformers for image recognition at scale. In International Conference on Learning Representations, 2021.
  • French (1999) French, R. M. Catastrophic forgetting in connectionist networks. Trends in cognitive sciences, 3(4):128–135, 1999.
  • Gu et al. (2022) Gu, Y., Yang, X., Wei, K., and Deng, C. Not Just Selection, but Exploration: Online Class-Incremental Continual Learning via Dual View Consistency. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  7432–7441, 2022.
  • Guo et al. (2023) Guo, Y., Liu, B., and Zhao, D. Dealing with cross-task class discrimination in online continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  11878–11887, 2023.
  • He (2024) He, J. Gradient reweighting: Towards imbalanced class-incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  16668–16677, 2024.
  • He et al. (2020) He, J., Mao, R., Shao, Z., and Zhu, F. Incremental learning in online scenario. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, June 2020.
  • He et al. (2016) He, K., Zhang, X., Ren, S., and Sun, J. Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp.  770–778, 2016.
  • Hendrycks et al. (2021) Hendrycks, D., Basart, S., Mu, N., Kadavath, S., Wang, F., Dorundo, E., Desai, R., Zhu, T., Parajuli, S., Guo, M., et al. The many faces of robustness: A critical analysis of out-of-distribution generalization. In Proceedings of the IEEE/CVF international conference on computer vision, pp.  8340–8349, 2021.
  • Hsu et al. (2018) Hsu, Y.-C., Liu, Y.-C., Ramasamy, A., and Kira, Z. Re-evaluating continual learning scenarios: A categorization and case for strong baselines. arXiv preprint arXiv:1810.12488, 2018.
  • Keskar & Socher (2017) Keskar, N. S. and Socher, R. Improving generalization performance by switching from adam to sgd. arXiv preprint arXiv:1712.07628, 2017.
  • Kingma & Ba (2014) Kingma, D. P. and Ba, J. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  • Kirkpatrick et al. (2017) Kirkpatrick, J., Pascanu, R., Rabinowitz, N., Veness, J., Desjardins, G., Rusu, A. A., Milan, K., Quan, J., Ramalho, T., Grabska-Barwinska, A., et al. Overcoming catastrophic forgetting in neural networks. Proceedings of the national academy of sciences, 114(13):3521–3526, 2017.
  • Koh et al. (2023) Koh, H., Seo, M., Bang, J., Song, H., Hong, D., Park, S., Ha, J.-W., and Choi, J. Online boundary-free continual learning by scheduled data prior. In International Conference on Learning Representations, 2023.
  • Krizhevsky (2012) Krizhevsky, A. Learning multiple layers of features from tiny images. University of Toronto, 05 2012.
  • Lin et al. (2023) Lin, H., Zhang, B., Feng, S., Li, X., and Ye, Y. Pcr: Proxy-based contrastive replay for online class-incremental continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  24246–24255, 2023.
  • Mai et al. (2021) Mai, Z., Li, R., Kim, H., and Sanner, S. Supervised contrastive replay: Revisiting the nearest class mean classifier in online class-incremental continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  3589–3599, 2021.
  • Mai et al. (2022) Mai, Z., Li, R., Jeong, J., Quispe, D., Kim, H., and Sanner, S. Online continual learning in image classification: An empirical survey. Neurocomputing, 469:28–51, 2022.
  • McDonnell et al. (2024) McDonnell, M. D., Gong, D., Parvaneh, A., Abbasnejad, E., and van den Hengel, A. Ranpac: Random projections and pre-trained models for continual learning. Advances in Neural Information Processing Systems, 36, 2024.
  • Michel et al. (2024) Michel, N., Wang, M., Xiao, L., and Yamasaki, T. Rethinking momentum knowledge distillation in online continual learning. In Forty-first International Conference on Machine Learning, 2024.
  • Mirzadeh et al. (2020) Mirzadeh, S. I., Farajtabar, M., Pascanu, R., and Ghasemzadeh, H. Understanding the role of training regimes in continual learning. Advances in Neural Information Processing Systems, 33:7308–7320, 2020.
  • Moon et al. (2023) Moon, J.-Y., Park, K.-H., Kim, J. U., and Park, G.-M. Online class incremental learning on stochastic blurry task boundary via mask and visual prompt tuning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp.  11731–11741, 2023.
  • Paszke et al. (2019) Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., et al. Pytorch: An imperative style, high-performance deep learning library. Advances in Neural Information Processing Systems, 32, 2019.
  • Rao et al. (2019) Rao, D., Visin, F., Rusu, A., Pascanu, R., Teh, Y. W., and Hadsell, R. Continual unsupervised representation learning. Advances in neural information processing systems, 32, 2019.
  • Roy et al. (2024) Roy, A., Moulick, R., Verma, V. K., Ghosh, S., and Das, A. Convolutional prompting meets language models for continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  23616–23626, 2024.
  • Ruder (2016) Ruder, S. An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747, 2016.
  • Smith et al. (2023) Smith, J. S., Karlinsky, L., Gutta, V., Cascante-Bonilla, P., Kim, D., Arbelle, A., Panda, R., Feris, R., and Kira, Z. Coda-prompt: Continual decomposed attention-based prompting for rehearsal-free continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  11909–11919, 2023.
  • Tiwari et al. (2022) Tiwari, R., Killamsetty, K., Iyer, R., and Shenoy, P. Gcr: Gradient coreset based replay buffer selection for continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  99–108, 2022.
  • Vaswani (2017) Vaswani, A. Attention is all you need. Advances in Neural Information Processing Systems, 2017.
  • Wah et al. (2011) Wah, C., Branson, S., Welinder, P., Perona, P., and Belongie, S. The caltech-ucsd birds-200-2011 dataset. Technical Report CNS-TR-2011-001, California Institute of Technology, 2011.
  • Wang et al. (2024) Wang, M., Michel, N., Xiao, L., and Yamasaki, T. Improving plasticity in online continual learning via collaborative learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  23460–23469, 2024.
  • Wang et al. (2022a) Wang, Z., Zhang, Z., Ebrahimi, S., Sun, R., Zhang, H., Lee, C.-Y., Ren, X., Su, G., Perot, V., Dy, J., et al. Dualprompt: Complementary prompting for rehearsal-free continual learning. In European Conference on Computer Vision, pp.  631–648. Springer, 2022a.
  • Wang et al. (2022b) Wang, Z., Zhang, Z., Lee, C.-Y., Zhang, H., Sun, R., Ren, X., Su, G., Perot, V., Dy, J., and Pfister, T. Learning to prompt for continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  139–149, 2022b.
  • Zhou et al. (2024a) Zhou, D.-W., Sun, H.-L., Ning, J., Ye, H.-J., and Zhan, D.-C. Continual learning with pre-trained models: A survey. In International Joint Conference on Artificial Intelligence, pp.  8363–8371, 2024a.
  • Zhou et al. (2024b) Zhou, D.-W., Sun, H.-L., Ye, H.-J., and Zhan, D.-C. Expandable subspace ensemble for pre-trained model-based class-incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  23554–23564, 2024b.

Appendix A Algorithm of our Adam Implementation

As explained in Section 4.2, the actual implementation that we used for our experiments is based on an Adam update. For the sake of clarity, we presented our method with SGD. Similarly, we omitted the bias from the pseudo-code. Therefore, we give the full details of the procedure in Algorithm 2, in a pseudo-code Pytorch-like implementation.

# Adam parameters
m = 0
v = 0
beta1 = 0.9
beta2 = 0.999
step = 0
\par# Hypergrad parameters
gamma = 1e-3
grad_weight = torch.ones(n_classes)
prev_grad = None
for x, y in dataloader:
# Baseline loss
h, logits_base = network(x) # features and logits
# Batch-wise masking
logits_base[:, [i for i in range(logits_base.shape[-1]) if i not in y.unique()]] = float(’-inf’)
loss_baseline = criterion_baseline(logits_base, y)
\par# FC recalibration
proto, labels = get_prototypes()
logits = network.fc(proto)
# Batch-wise masking
logits[:, [i for i in range(logits.shape[-1]) if i not in label.unique()]] = float(’-inf’)
loss_op = cross_entropy(logits, labels) # Eq. 10
\parloss = loss_baseline + loss_op # Eq. 11
\paroptim.zero_grad()
loss.backward()
\par# Class-Wise Hypergradient update
curr_W = network.fc.weight.grad
curr_B = network.fc.bias.grad
curr_grad = torch.cat([curr_W, curr_B.unsqueeze(1)], dim=1)
if prev_grad is not None:
# Adam update
m = beta1 * m + (1 - beta1) * curr_grad
v = beta2 * v + (1 - beta2) * (curr_grad ** 2)
m_hat = m / (1 - beta1 ** step)
v_hat = v / (1 - beta2 ** step)
curr_grad = m_hat / (torch.sqrt(v_hat) + 1e-8)
\pargrad_weight += gamma * (curr_grad @ prev_grad.T).diag() #Eq. 8
for i in range(n_classes):
network.fc.weight.grad[i, :] = network.fc.weight.grad[i, :] * grad_weight[i]
network.fc.bias.grad[i] = network.fc.bias.grad[i] * grad_weight[i]
prev_grad = curr_grad
optim.step()
\parupdate_proto(h, y) # Eq. 9
Algorithm 2 PyTorch-like pseudo-code of our Adam-based method integration with other baselines. Extra details are given in this version regarding bias consideration and batch-wise masking.

Appendix B Datasets and Baselines

B.1 Datasets

PTMs are often pre-trained on ImageNet-21k, making it unfair to experiment on such datasets. To showcase the performances of our approach we experiment with the following:

  • CUB (Wah et al., 2011): The CUB dataset (Caltech-UCSD Birds-200) contains 200 bird species with 11,788 images, annotated with attributes and part locations for fine-grained classification.

  • ImageNet-R (Hendrycks et al., 2021): ImageNet-R is a set of images labeled with ImageNet label renditions. It contains 30,000 images spanning 200 classes, focusing on robustness with images in various artistic styles.

  • CIFAR100 (Krizhevsky, 2012): CIFAR-100 consists of 60,000 32x32 color images across 100 classes, with 500 images per class, split into 500 training and 100 test samples per class.

B.2 Baselines

Prompt learning-based methods (Zhou et al., 2024a) are particularly suited for being combined with our approach in onCL as they all capitalize on a final FC layer for classification. Therefore, we consider the following.

  • L2P (Wang et al., 2022b): Learning to Prompt (L2P) is the foundation of prompt learning methods in Continual Learning. The main idea is to learn how to append a fixed-sized prompt to the input of the ViT (Dosovitskiy et al., 2021). The ViT stays frozen, only the appended prompt as well as the FC layer are trained.

  • DualPrompt (Wang et al., 2022a): DualPrompt follows closely the work of L2P by addressing forgetting in the prompt level with task-specific prompts as well as higher lever long-term prompts.

  • CODA (Smith et al., 2023): CODA-prompt improves prompt learning by computing prompt on the fly leveraging a component pool and an attention mechanism. Therefore, CODA benefits from a single gradient flow and achieves state-of-the-art performances.

  • ConvPrompt (Roy et al., 2024): ConvPrompt leverages convolutional prompts and dynamic task-specific embeddings while incorporating text information from large language models.

  • MVP (Moon et al., 2023): MVP uses instance-wise logit masking, contrastive visual prompt tuning for Continual Learning in the Si-Blurry context.

Appendix C Additionnal Evaluation Metrics

Here we report additional metrics in the clear and blurry contexts for all methods for additional insights into the performances

C.1 Final Average Accuracy

We report the final Average Accuracy 𝒜T\mathcal{A}_{T} as per the definition of Section 5.1. Such results are presented in Table 8 and Table 9.

C.2 Average Performances on Old Classes

We report the Average Accuracy on old classes only at the end of training, 𝒜T1\mathcal{A}_{T-1} as per the definition of Section 5.1. Such results are presented in Table 8 and Table 9. A higher value of performance in old classes indicates a better ability to retain knowledge, also known as stability.

C.3 Additional Ablation

We include additional experiments with CODA on CUB to evaluate the impact of combining CWH only. Such results are presented in Table 6. As we can see, CODA still benefits the most from OP, however, it reaches the best performances in average across learning rate when being combined with OP and CWH.

Table 6: AP on CUB for CODA in the Si-Blurry setting, with and without CWH, with and without OP. We report the average and standard deviation over 5 runs.
Dataset CUB
Learning Rate 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg.
CODA 12.08±1.31 58.4±1.45 44.42±6.92 38.30±23.75
     \hookrightarrow + CWH 17.08±1.5 59.17±2.29 43.52±3.73 39.93±21.27
     \hookrightarrow + OP 37.31±1.53 77.47±1.36 59.78±4.17 58.18±20.13
     \hookrightarrow + CWH + OP 50.01±1.35 78.49±1.43 48.23±9.24 58.91±16.99

C.4 Time Complexity

Experiments were run on various machines including Quadro RTX 8000 GPU, Tesla V100 16Go GPU, A100 40Go. In this section, we report the times of execution of each method, as well as the overhead induced by leveraging our components. To do so we run all methods on a single V100-16Go. The results are presented in Table 7. As expected, the time consumption overhead of including OP and CWH is minimal.

Table 7: Time consumption of all methods with and without our components, on CIFAR100, clear setup, on a single V100-16Go.
Component MVP DualPrompt CODA ConvPrompt L2P
Baseline 24m 11s 21m 10s 23m 6s 2h 32m 15s 20m 7s
+ CWH 25m 8s 22m 7s 23m 22s 2h 32m 43s 20m 28s
+ OP 25m 25s 22m 36s 24m 23s 2h 33m 31s 21m 27s
+ CWH + OP 26m 44s 23m 7s 24m 40s 2h 34m 40s 21m 33s

C.5 Spatial Complexity

Class-Wise Hypergradients.

The usage of CWH solely requires storing one float per class (with cc classes total) as well as previous gradient values in the last FC layer in the case of SGD. This amounts to a total of c×c×(l+1)c\times c\times(l+1) additional floats to store. We multiply by l+1l+1 to account for the bias. For CIFAR100 a Vit-base, we have c=100c=100 and l=768l=768. In the case of Adam update, Adam parameters must equally be included.

Online Prototypes.

Storing OP only requires one vector of dimension ll per class, with l=768l=768 in the case of ViT base. Additionally, an extra integer per class must be stored to keep track of the index of the update of each OP. If the index is stored as a float, the additional amount of floating points to store is c×(l+1)c\times(l+1), with cc the number of classes, and ll the output dimension of the PTM.

Appendix D Adaptation of Methods to our setup

Since most methods compared here were originally designed for offCL, they had to be specifically adapted to the onCL scenario. In that sense, some parameters have been chosen arbitrarily, based on their offCL values, without additional hyper-parameter search. Such a situation is similar to one that would be observed in real-world cases where an offCL model tries to be adapted to an onCL problem. For all methods, we use a fixed initial learning rate, no scheduler, and Adam optimizer. Of course, we disabled an operation that was operated at task change. Additionally, even though MVP was indeed designed for online cases, we found several differences between their training procedure and ours, which we discuss below.

Adaptation of CODA.

In their original paper and implementation, the authors require freezing components after each task, therefore having task-specific components. Typically, they show that performances tend to plateau for more than 100100 components, and for a 1010-tasks sequence, they would reserve 1010 components per task. In our implementation, we decided to similarly use 100100 components for the entire training. However, we train all components together at all times during training since we cannot know when would be the correct time to freeze or unfreeze them. For other parameters, we followed the original implementation.

Adaptation of ConvPrompt.

ConvPrompt is a method that heavily relies on task boundaries in its original implementation, notably by incorporating 55 new prompts per task. Contrarily to CODA, allocating the maximum number of prompt generators at all times, without a freeze, would induce an important training time constraint. Therefore, we only use 55 prompt generators at all times. Despite this reduction in overall parameters, ConvPrompt still achieves competitive results in the clear setting. However, its performances drastically fall off in the Si-Blurry case. Further, an in-depth adaptation of ConvPrompt in the online context could potentially improve its performance, however, such a study is not covered in this work.

Another interesting aspect of ConvPrompt adaptation into our setup is the high variance observed when the LR is high (5×1035\times 10^{-3}). The reason for such results is that ConvPrompt is indeed very sensitive to the choice of the initial LR, therefore, when is it too large the training is unstable and some runs result in extremely poor performances. This observation once again highlights the importance of properly selecting the initial learning rate in onCL.

Adaptation of DualPrompt.

Similar to CODA, but on a prompt level, DualPrompt requires freezing prompts at task change. For adapting it to onCL, we chose to use all prompts at all times without freezing the prompt pool. The E-Prompt pool size is set to 1010 and the G-Prompt pool size is set to 55.

Adaptation of L2P.

The same logic as the one described for CODA and DualPrompt applies to L2P. In that sense, we use the entire prompt pool at all times without freezing. The prompt pool size is set to 1010.

Adaptation and Performances of MVP

Even though MVP is designed for the online case, their original training setup differs slightly. Firstly, the batch size is set to 3232 (we use 1010), and they similarly consider that each batch can be used for 3 separate gradient steps. In that sense, the performances reported in the original paper might be higher as they trained on a slightly more constrained setup. Secondly, the authors use the same learning rate and optimizer for each compared method, which as we argued in this work, might lead to different results relatively speaking compared to other methods. Such experimental differences might lead to the performances obtained in our experiments which are in most cases significantly lower than other methods.

Appendix E Details on the Si-Blurry Setting

We followed the procedure and code made available by the authors of MVP (Moon et al., 2023) in order to generate the Si-Blurry versions of the datasets. Notably, we use M=10M=10 and N=50N=50, following the original work. The number of tasks is the same as in the clear setting. We show the number of images per class apparition during training for a subset of classes to give a better overview of this training environment in Figure 5.

Refer to caption
Figure 5: Example of class apparition during training in the Si-Blurry setting. The y-axis represents the average number of images of a given class present in the current batch of size 1010.
Table 8: Final Average Accuracy (%) of all considered baselines with and without our proposed OP and CWH, in the clear setting. Results are displayed for different datasets and LR values. We also report the average across LR. We used 3 runs and reported the average and standard deviations. Best viewed in color.
Dataset CIFAR100 Imagenet-R CUB
Learning Rate 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg. 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg. 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg.
L2P 66.88±0.48 69.75±1.75 59.74±0.74 65.46±0.99 45.16±1.13 47.24±1.86 26.68±2.31 39.69±1.77 36.39±3.12 39.2±1.75 17.81±1.11 31.13±1.99
     \hookrightarrow + OP 72.19±0.11 79.54±0.39 73.98±0.64 75.24±0.38 52.81±0.71 63.42±0.29 46.13±0.19 54.12±0.4 60.67±2.3 75.11±2.46 46.95±2.51 60.91±2.42
        \hookrightarrow + CWH 73.33±0.47 79.36±0.32 74.92±0.7 75.87±0.5 52.46±1.09 63.15±0.28 49.17±0.35 54.93±0.57 60.25±3.23 75.06±1.8 56.4±1.52 63.9±2.18
DualPrompt 62.16±1.59 69.41±1.33 54.57±0.68 62.05±1.2 40.94±6.68 51.06±1.14 24.5±1.17 38.83±3.0 35.07±3.39 47.61±2.84 21.77±0.71 34.82±2.31
     \hookrightarrow + OP 67.04±2.75 79.9±0.25 72.06±1.01 73.0±1.34 40.77±4.7 61.84±0.42 48.46±0.59 50.36±1.9 52.53±3.32 78.53±0.61 57.34±0.7 62.8±1.54
        \hookrightarrow + CWH 64.62±2.29 79.43±0.48 73.2±1.07 72.42±1.28 43.99±6.43 61.25±0.27 52.08±0.64 52.44±2.45 51.01±5.05 78.89±0.94 65.7±0.66 65.2±2.22
CODA 61.63±3.44 75.25±1.87 67.68±1.08 68.19±2.13 26.51±1.21 55.25±2.49 27.56±0.97 36.44±1.56 20.82±12.31 47.92±1.96 22.56±0.52 30.43±4.93
     \hookrightarrow + OP 67.47±4.36 79.53±0.72 82.01±0.28 76.34±1.79 35.44±4.43 64.76±5.24 54.87±0.78 51.69±3.48 40.98±12.96 73.73±2.23 55.87±0.47 56.86±5.22
        \hookrightarrow + CWH 69.1±0.43 80.13±1.85 82.89±0.41 77.37±0.9 33.61±2.56 61.62±4.91 58.93±0.84 51.39±2.77 43.91±6.46 74.2±1.83 64.8±0.43 60.97±2.91
ConvPrompt 69.07±0.69 79.13±0.97 70.83±0.95 73.01±0.87 14.48±24.17 59.45±2.73 32.26±2.92 35.4±9.94 0.4±0.09 56.56±2.67 29.66±0.22 28.87±0.99
     \hookrightarrow + OP 77.45±2.63 86.14±0.26 84.12±0.57 82.57±1.15 16.99±29.14 70.15±0.81 57.69±0.98 48.28±10.31 0.4±0.11 79.58±1.78 57.38±1.56 45.79±1.15
        \hookrightarrow + CWH 79.89±3.74 85.69±0.41 84.78±0.64 83.45±1.6 33.93±28.98 66.77±2.36 57.38±1.03 52.69±10.79 61.89±13.97 75.64±3.33 66.29±0.85 67.94±6.05
MVP 27.37±1.99 40.58±1.61 33.85±5.76 33.93±3.12 21.61±1.2 30.12±0.24 27.23±0.66 26.32±0.7 13.02±1.74 28.58±2.76 26.36±0.53 22.65±1.68
     \hookrightarrow + OP 46.0±2.88 80.0±0.47 75.83±3.94 67.28±2.43 26.63±2.05 55.48±0.91 55.83±0.29 45.98±1.08 40.77±1.59 80.13±0.53 71.78±1.34 64.23±1.15
        \hookrightarrow + CWH 51.06±2.45 79.82±0.2 76.61±3.43 69.16±2.03 33.52±5.59 55.56±0.97 56.53±0.33 48.54±2.3 44.84±4.68 79.19±0.13 74.03±1.0 66.02±1.94
Table 9: Final Average Accuracy (%) of all considered baselines with and without our proposed OP and CWH, in the Si-Blurry setting. Results are displayed for different datasets and LR values. We also report the average across LR. We used 3 runs and reported the average and standard deviations. Best viewed in color.
Dataset CIFAR100 Imagenet-R CUB
Learning Rate 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg. 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg. 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg.
L2P 65.74±0.46 70.09±1.44 62.66±3.44 66.16±1.78 46.39±0.84 38.78±14.14 29.02±1.39 38.06±5.46 40.09±0.78 41.25±0.93 23.24±1.1 34.86±0.94
     \hookrightarrow + OP 74.77±0.29 80.15±1.05 74.0±0.77 76.31±0.7 52.9±2.59 62.78±0.6 47.12±1.22 54.27±1.47 60.54±2.97 74.97±0.83 48.95±0.85 61.49±1.55
        \hookrightarrow + CWH 75.16±0.91 79.68±1.13 74.74±0.64 76.53±0.89 51.18±0.53 62.37±0.54 49.92±0.71 54.49±0.59 56.9±4.2 74.97±1.02 58.34±0.26 63.4±1.83
DualPrompt 62.23±0.55 71.36±0.73 57.53±0.36 63.71±0.55 41.69±7.76 50.31±2.65 26.86±1.11 39.62±3.84 39.36±3.86 49.29±3.72 25.74±1.58 38.13±3.05
     \hookrightarrow + OP 70.49±2.71 80.11±0.21 72.68±1.98 74.43±1.63 43.2±5.4 62.07±0.6 50.17±1.4 51.81±2.47 56.67±4.01 78.92±1.08 59.76±1.12 65.12±2.07
        \hookrightarrow + CWH 70.0±1.64 79.46±0.34 73.58±1.68 74.35±1.22 39.75±5.28 60.92±0.36 53.54±1.32 51.4±2.32 55.87±2.7 79.3±1.3 68.04±1.48 67.74±1.83
CODA 61.87±1.35 75.78±1.11 69.54±0.14 69.06±0.87 27.87±9.2 58.1±0.18 31.14±1.26 39.04±3.55 38.86±4.27 50.82±2.12 28.99±1.01 39.56±2.47
     \hookrightarrow + OP 74.47±1.39 82.53±1.69 82.59±0.33 79.86±1.14 31.5±11.37 66.45±0.64 56.67±1.13 51.54±4.38 50.88±6.34 74.68±2.02 60.91±0.02 62.16±2.79
        \hookrightarrow + CWH 74.03±0.74 82.5±1.76 83.36±0.39 79.96±0.96 37.6±0.97 63.65±3.45 59.69±1.04 53.65±1.82 50.95±4.63 74.06±1.17 68.8±0.94 64.6±2.25
ConvPrompt 45.37±38.07 80.02±0.39 74.57±1.54 66.65±13.33 12.72±21.2 59.98±1.18 35.48±2.65 36.06±8.34 37.63±2.61 54.71±5.95 34.51±3.02 42.28±3.86
     \hookrightarrow + OP 79.37±0.7 87.07±0.89 84.48±1.24 83.64±0.94 31.29±27.08 71.08±0.9 58.49±3.31 53.62±10.43 8.2±13.32 79.93±1.42 64.74±0.68 50.96±5.14
        \hookrightarrow + CWH 80.97±1.61 86.67±1.09 84.16±1.51 83.93±1.4 35.68±29.53 69.36±1.03 56.97±3.83 54.0±11.46 67.4±3.0 78.17±6.07 70.64±1.17 72.07±3.41
MVP 32.8±6.71 48.33±3.44 40.54±2.18 40.56±4.11 23.25±4.73 31.64±0.87 29.04±0.99 27.98±2.2 19.07±2.05 33.45±1.62 27.89±1.44 26.8±1.7
     \hookrightarrow + OP 49.37±3.54 81.38±0.63 76.62±0.64 69.12±1.6 27.99±11.28 57.53±0.4 56.56±0.24 47.36±3.97 44.53±6.4 80.66±0.71 73.64±1.23 66.28±2.78
        \hookrightarrow + CWH 52.51±4.18 81.07±0.71 77.11±0.7 70.23±1.86 33.84±2.58 57.47±0.65 57.18±0.36 49.5±1.2 45.7±11.3 80.51±2.1 76.32±1.5 67.51±4.97
Table 10: Average Accuracy on Old Classes (%) of all considered baselines with and without our proposed OP and CWH, in the clear setting. Results are displayed for different datasets and LR values. We also report the average across LR. We used 3 runs and reported the average and standard deviations. Best viewed in color.
Dataset CIFAR100 Imagenet-R CUB
Learning Rate 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg. 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg. 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg.
L2P 64.81±0.78 68.49±2.81 58.48±1.74 63.93±1.78 44.21±1.3 46.78±1.96 26.29±1.8 39.09±1.69 34.69±2.71 38.0±1.24 17.76±0.18 30.15±1.38
     \hookrightarrow + OP 70.16±0.3 79.05±1.15 75.6±1.17 74.94±0.87 51.88±0.81 63.49±0.19 47.51±0.12 54.29±0.37 59.48±2.8 75.42±1.74 48.84±2.41 61.25±2.32
        \hookrightarrow + CWH 71.21±0.82 78.74±1.07 76.29±1.23 75.41±1.04 51.51±1.24 63.19±0.46 50.57±0.23 55.09±0.64 59.27±3.19 75.66±1.22 58.99±1.32 64.64±1.91
DualPrompt 59.46±1.03 67.76±1.84 53.06±1.24 60.09±1.37 39.3±7.18 50.7±0.94 24.42±1.15 38.14±3.09 33.27±3.24 46.71±2.59 21.12±0.94 33.7±2.26
     \hookrightarrow + OP 64.64±3.07 79.69±0.4 73.89±1.21 72.74±1.56 39.04±4.91 61.83±0.25 50.11±0.67 50.33±1.94 50.85±3.05 78.56±0.47 59.21±0.96 62.87±1.49
        \hookrightarrow + CWH 61.75±2.5 78.99±0.47 74.83±1.18 71.86±1.38 42.41±6.84 61.14±0.34 53.62±0.65 52.39±2.61 49.65±5.22 79.04±0.7 67.91±1.08 65.53±2.33
CODA 59.93±2.72 73.64±2.47 65.72±1.64 66.43±2.28 25.42±1.26 54.4±2.67 26.85±0.98 35.56±1.64 18.88±12.78 46.51±1.9 21.84±0.65 29.08±5.11
     \hookrightarrow + OP 64.93±4.38 78.08±1.16 81.58±0.37 74.86±1.97 34.38±4.77 64.21±5.79 56.14±0.52 51.58±3.69 39.67±13.24 73.15±2.29 57.89±0.61 56.9±5.38
        \hookrightarrow + CWH 66.42±0.3 78.8±2.36 82.56±0.8 75.93±1.15 31.61±3.21 61.02±5.09 60.07±0.6 50.9±2.97 42.06±7.08 73.91±2.31 67.19±0.98 61.05±3.46
ConvPrompt 67.2±0.46 77.77±1.07 69.11±0.44 71.36±0.66 14.03±23.53 58.72±2.96 31.61±2.77 34.79±9.75 0.31±0.27 55.56±2.51 29.27±1.17 28.38±1.32
     \hookrightarrow + OP 75.96±3.12 85.49±0.57 83.37±0.61 81.61±1.43 16.58±28.41 69.89±1.09 58.13±0.99 48.2±10.16 0.1±0.18 79.18±1.82 59.03±1.76 46.1±1.25
        \hookrightarrow + CWH 78.54±4.24 84.88±0.63 84.26±0.44 82.56±1.77 32.92±28.63 66.31±2.56 58.6±1.13 52.61±10.77 60.58±14.61 75.0±3.08 69.01±0.87 68.2±6.19
MVP 19.82±1.99 34.39±1.73 27.47±6.2 27.23±3.31 18.19±1.73 26.99±0.9 25.3±0.72 23.49±1.12 8.74±1.9 24.91±2.98 24.36±1.16 19.34±2.01
     \hookrightarrow + OP 40.51±3.49 78.72±1.0 75.12±4.79 64.78±3.09 23.42±1.99 53.73±1.23 55.23±0.49 44.13±1.24 37.8±2.03 79.27±0.71 71.69±1.99 62.92±1.58
        \hookrightarrow + CWH 46.15±2.57 78.54±0.81 75.81±4.27 66.83±2.55 30.81±6.23 53.82±1.22 55.62±0.53 46.75±2.66 42.18±4.9 78.28±0.11 73.5±1.51 64.65±2.17
Table 11: Average Accuracy on Old Classes (%) of all considered baselines with and without our proposed OP and CWH, in the Si-Blurry setting. Results are displayed for different datasets and LR values. We also report the average across LR. We used 3 runs and reported the average and standard deviations. Best viewed in color.
Dataset CIFAR100 Imagenet-R CUB
Learning Rate 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg. 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg. 5×1055\times 10^{-5} 5×1045\times 10^{-4} 5×1035\times 10^{-3} Avg.
L2P 65.74±0.46 70.09±1.44 62.66±3.44 66.16±1.78 46.39±0.84 38.78±14.14 29.02±1.39 38.06±5.46 40.09±0.78 41.25±0.93 23.24±1.1 34.86±0.94
     \hookrightarrow + OP 74.77±0.29 80.15±1.05 74.0±0.77 76.31±0.7 52.9±2.59 62.78±0.6 47.12±1.22 54.27±1.47 60.54±2.97 74.97±0.83 48.95±0.85 61.49±1.55
        \hookrightarrow + CWH 75.16±0.91 79.68±1.13 74.74±0.64 76.53±0.89 51.18±0.53 62.37±0.54 49.92±0.71 54.49±0.59 56.9±4.2 74.97±1.02 58.34±0.26 63.4±1.83
DualPrompt 62.23±0.55 71.36±0.73 57.53±0.36 63.71±0.55 41.69±7.76 50.31±2.65 26.86±1.11 39.62±3.84 39.36±3.86 49.29±3.72 25.74±1.58 38.13±3.05
     \hookrightarrow + OP 70.49±2.71 80.11±0.21 72.68±1.98 74.43±1.63 43.2±5.4 62.07±0.6 50.17±1.4 51.81±2.47 56.67±4.01 78.92±1.08 59.76±1.12 65.12±2.07
        \hookrightarrow + CWH 70.0±1.64 79.46±0.34 73.58±1.68 74.35±1.22 39.75±5.28 60.92±0.36 53.54±1.32 51.4±2.32 55.87±2.7 79.3±1.3 68.04±1.48 67.74±1.83
CODA 61.87±1.35 75.78±1.11 69.54±0.14 69.06±0.87 27.87±9.2 58.1±0.18 31.14±1.26 39.04±3.55 38.86±4.27 50.82±2.12 28.99±1.01 39.56±2.47
     \hookrightarrow + OP 74.47±1.39 82.53±1.69 82.59±0.33 79.86±1.14 31.5±11.37 66.45±0.64 56.67±1.13 51.54±4.38 50.88±6.34 74.68±2.02 60.91±0.02 62.16±2.79
        \hookrightarrow + CWH 74.03±0.74 82.5±1.76 83.36±0.39 79.96±0.96 37.6±0.97 63.65±3.45 59.69±1.04 53.65±1.82 50.95±4.63 74.06±1.17 68.8±0.94 64.6±2.25
ConvPrompt 45.37±38.07 80.02±0.39 74.57±1.54 66.65±13.33 12.72±21.2 59.98±1.18 35.48±2.65 36.06±8.34 37.63±2.61 54.71±5.95 34.51±3.02 42.28±3.86
     \hookrightarrow + OP 79.37±0.7 87.07±0.89 84.48±1.24 83.64±0.94 31.29±27.08 71.08±0.9 58.49±3.31 53.62±10.43 8.2±13.32 79.93±1.42 64.74±0.68 50.96±5.14
        \hookrightarrow + CWH 80.97±1.61 86.67±1.09 84.16±1.51 83.93±1.4 35.68±29.53 69.36±1.03 56.97±3.83 54.0±11.46 67.4±3.0 78.17±6.07 70.64±1.17 72.07±3.41
MVP 32.8±6.71 48.33±3.44 40.54±2.18 40.56±4.11 23.25±4.73 31.64±0.87 29.04±0.99 27.98±2.2 19.07±2.05 33.45±1.62 27.89±1.44 26.8±1.7
     \hookrightarrow + OP 49.37±3.54 81.38±0.63 76.62±0.64 69.12±1.6 27.99±11.28 57.53±0.4 56.56±0.24 47.36±3.97 44.53±6.4 80.66±0.71 73.64±1.23 66.28±2.78
        \hookrightarrow + CWH 52.51±4.18 81.07±0.71 77.11±0.7 70.23±1.86 33.84±2.58 57.47±0.65 57.18±0.36 49.5±1.2 45.7±11.3 80.51±2.1 76.32±1.5 67.51±4.97