This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Continual Learning with Dependency Preserving Hypernetworks

Dupati Srikar Chandra1 Sakshi Varshney1 P.K. Srijith1 Sunil Gupta2
1Indian Institute of Technology, India 2Deakin University, Australia
{ai20resch11004,cs16resch01002}@iith.ac.in srijith@cse.iith.ac.in sunil.gupta@deakin.edu.au
Abstract

Humans learn continually throughout their lifespan by accumulating diverse knowledge and fine-tuning it for future tasks. When presented with a similar goal, neural networks suffer from catastrophic forgetting if data distributions across sequential tasks are not stationary over the course of learning. An effective approach to address such continual learning (CL) problems is to use hypernetworks which generate task dependent weights for a target network. However, the continual learning performance of existing hypernetwork based approaches are affected by the assumption of independence of the weights across the layers in order to maintain parameter efficiency. To address this limitation, we propose a novel approach that uses a dependency preserving hypernetwork to generate weights for the target network while also maintaining the parameter efficiency. We propose to use recurrent neural network (RNN) based hypernetwork that can generate layer weights efficiently while allowing for dependencies across them. In addition, we propose novel regularisation and network growth techniques for the RNN based hypernetwork to further improve the continual learning performance. To demonstrate the effectiveness of the proposed methods, we conducted experiments on several image classification continual learning tasks and settings. We found that the proposed methods based on the RNN hypernetworks outperformed the baselines in all these CL settings and tasks.

1 Introduction

There are various applications where a computational system has to learn from a stream of data continually and adapt to the environment by using knowledge gained from past experiences. For example, Autonomous agents in the real world have to learn over continuous streams of data and need to remember the information from various non-stationary distributions without forgetting [20]. Deep neural networks have achieved high performance on many images classification benchmarks, comparable or even better than humans. However, while learning a new task, these networks forget the knowledge gained from previous ones. The process of forgetting knowledge or information gained from previous tasks due to drastic weight alteration while learning new tasks is known as catastrophic forgetting [18]. If data distributions across the sequential tasks is not stationary, the weights of the model alter drastically to classify the new task, leading to forgetting of knowledge gained from previous tasks. Ideally, performance of a newly learnt task should not have an impact on previous one (or vice versa). To overcome forgetting, computational systems or agents, on one hand, should be plastic to acquire new information and refine old information based on continuous input and, on the other hand, it should be stable to prevent the novel input from interfering with old information. This is referred to as plasticity-stability dilemma [7, 19]. Continual learning aims to develop machine learning and deep learning models which are stable enough to retain information learnt from old tasks but also has the required plasticity to learn new tasks [20].

Continual learning techniques for neural networks has gained significant attention recently [20]. Several continual learning methods have been proposed to avoid forgetting in neural networks. These can be broadly categorised into three approaches, Regularisation techniques [9, 12], dynamic architecture methods [2, 24] and replay based methods  [5, 21]. Recently, Continual learning with hypernetworks has shown very promising results in dealing with forgetting at a meta level by generating task specific weights [4, 8, 28]. Hypernetwork is a meta neural network which generates parameters for the main network associated with the task (for instance, classification or regression network) by considering some task related information. During training, instead of directly trying to update parameters of the main network, hypernetwork parameters which generates them are updated. But generating the entire main network parameters will require a larger hypernetwork as it requires generating a very high dimensional output. Instead, chunked hypernetworks [8, 28] generate them in smaller chunks (chunk referred to subset of main network weights) multiple times iteratively using smaller hypernetwork that is reusable and also help in model compression significantly. A key limitation of chunked hypernetworks is that it assumes the weight matrices associated with the chunks to be independent, which affects the continual learning performance significantly. Chunked hypernetwork [28] uses a feed forward neural network to generate weights without considering dependencies across chunks. We propose to use recurrent neural networks (RNNs), in order to capture dependencies in weight matrix generation across the chunks. But standard RNNs suffer from vanishing gradient problems and may not be able to remember dependencies for a long time. So, a variant of RNN, LSTM [9] has been used to remember the weight matrix dependencies over a longer duration. Thus, we propose LSTM based hypernetwork that can efficiently generate weights of the main network while also maintaining the dependencies across the chunks.

While learning multiple tasks in sequential manner, the hypernetwork should remember the knowledge gained from previous tasks and should also be able to forward that knowledge to the upcoming tasks. To achieve this, we propose a novel hypernetwork regularisation technique, Importance Weighted Regularisation (IWR) that can further improve the performance of hypernetwork based continual learning by enabling forward transfer while also retaining previous information. IWR considers the importance of the main network parameters and allows more flexibility to the hypernetwork to adapt to the new task by considering this importance. We also propose a network growth technique for the LSTM based hypernetworks for continual learning. The approach is based on the idea that the dependence among the main network parameters across tasks remains the same while their exact values could be different. This is achieved by sharing the weights associated with the hidden states in the LSTM across the tasks. We still learn the input specific weights for each task separately. This approach improves the performance of continual learning with no extra regularisation and accelerates the model training. Our experimental results on real world data show that the proposed approaches along with LSTM based hypernetwork can effectively mitigate catastrophic forgetting and significantly improve the continual learning performance.

Our main contributions can be summarised as follows.

  • We propose a novel dependency preserving LSTM hypernetwork for continual learning.

  • We propose a novel regularisation technique for hypernetwork based continual learning and a network growth technique specifically for the LSTM based hypernetwork which does not require regularisation.

  • We demonstrate the improvement in continual learning performance of the proposed approaches through experiments on several image classification data sets and for different CL settings.

2 Related Work

Several approaches were proposed recently for continual learning and to deal with catastrophic forgetting. Conceptually, these approaches are classified based on replaying the stored examples, methods that expand the model on seeing the new task and methods that regularise the parameter shift by retaining the network [20, 13]. Regularisation based approaches [1, 11, 16, 30] avoid forgetting by imposing constraints on the update of parameters. With the advantage of no extra memory requirement, regularisation approaches are used for a broader variety of applications that have constraints on memory, computational resources, and data privacy. Elastic weight consolidation(EWC) [11] and Synaptic intelligence(SI) [30] are most well-known approaches, proposed to mitigate forgetting by constraining the update of important parameters of the task. It imposes a quadratic penalty on the difference between old and new parameters which helps to slow down the learning of new tasks by updating old parameters. But experiments in [10] showed that EWC is not effective in learning new classes incrementally.

Replay based methods [2, 17, 21] alleviate catastrophic forgetting by replaying old examples while learning new tasks. These methods either store examples from previous tasks or, generate synthetic examples from trained generative models from learnt feature space. Variational autoencoders (VAEs) and Generative adversarial networks (GANs) are used to generate samples from feature space. iCaRL [21], stores a subset of samples per class in fixed memory and selected examples should best approximate the class means in feature space. But as the number of tasks grows, samples per class stored will become too few to ensure required performance. Various other approaches [6, 22, 3] have been proposed which generate samples for old tasks instead of storing them and these are replayed while learning a new task.

Dynamic architecture based methods provide a solution towards continual learning by growing or updating its model structure for each task [15, 29]. Progressive neural network (PNNs) [23] grow their architecture by expanding the network statically with new modules. Forgetting can be circumvented by adding lateral connections from previous modules. Instead of growing the network structure statically, Dynamically expandable network (DEN) grows network architecture for each task with only a limited number of units and identifies neurons that are important for the new task and train them selectively [29].

Recently, hypernetwork based approaches have been proposed [8, 10, 26, 28] which has the advantage of having a constrained search space compared to the main network. Hypernetwork based techniques use a secondary neural network to generate the parameters of the main network and deal with forgetting at the hypernetwork level. To keep the size of the hypernetwork to be small compared to the main network, they generate the weights of the main network in small fixed sized chunks [8]. However, we notice that the process of independent generation of chunk weights ignores the dependence among main network parameters and thus affects the continual learning performance. To overcome this, we propose a LSTM based hypernetwork that can generate weights in smaller chunks while also maintaining dependency across them. Due to their ability to capture dependencies, LSTMs were used for learning from sequential data. Recently, continual learning methods for recurrent neural networks based on existing regularisation technique and hypernetworks were proposed in [4]. In contrast to the work in [4], where the aim is to model continual learning for tasks involving sequence data using RNNs with existing CL techniques, the goal of this paper is to develop novel continual learning methodology based on RNNs (specifically LSTMs) by treating them as a hypernetwork. In addition, we also introduce novel continual learning techniques for such LSTM based hypernetworks such as importance weighted regularisation and network growth.

3 LSTM Hypernetwork and Regularization Techniques for Continual Learning

In many realistic real world learning scenarios, the tasks arrive in sequential manner. Continual learning aims to learn from a sequence of tasks where the data of all the tasks are not available at once and we have a fixed memory size. We assume that we are given a sequence of KK tasks, where each task t𝒯={1,,K}t\in\mathcal{T}=\{1,\ldots,K\} contains input XtX^{t}={xjt}j=1nt\{x_{j}^{t}\}_{j=1}^{n_{t}} and target label YtY^{t}={yjt}j=1nt\{y_{j}^{t}\}_{j=1}^{n_{t}} where ntn_{t} being the number of samples in task t. The goal of main network mm is to learn a function fmt(,Θmt):XtYtf_{m}^{t}{(\cdot,{\Theta_{m}^{t}}}):X^{t}\to{Y^{t}} with parameters Θmt\Theta_{m}^{t} associated with task tt. While learning a task tt we have access to only observations of current task tt but no access to data of previous tasks.

We can learn Θmt\Theta_{m}^{t} separately for each task, but it results in a linear growth in number of parameters and the fixed sized memory will not be sufficient to store them. If we maintain the main network parameter to be same across all the tasks, the parameter values will get overwritten by the new task data which will results in catastrophic forgetting. To learn continuously over the tasks without requiring a linear growth in parameters, hypernetworks are proposed to generate the main network parameters for each task. The hypernetwork hh learns a function fh(,Θh):𝐞𝐭Θmtf_{h}{(\cdot,\Theta_{h}}):\mathbf{e^{t}}\to{\Theta_{m}^{t}} to generate the task specific parameter Θmt\Theta_{m}^{t} given a task embedding 𝐞𝐭\mathbf{e^{t}} using trainable parameters Θh\Theta_{h}.

Generating the high dimensional main network parameters all at once requires a very large hypernetwork with a large number of outputs and is computationally costly to train them. With the motive of reducing the number of trainable parameters in hypernetworks, chunked hypernetworks [8, 28] are proposed to generate weights in smaller chunks (subsets of the weight matrix of main network) by reusing the same hypernetwork fhf_{h} multiple times with different chunk embeddings.

Thus, hypernetwork fhf_{h} with parameters Θh\Theta_{h} takes task embedding 𝐞𝐭\mathbf{e^{t}} and chunk embeddings 𝐜\mathbf{c} = {𝐜1,,𝐜nc}\{\mathbf{c}_{1},\ldots,\mathbf{c}_{n_{c}}\} as input to generate set of main network weights, Θmt\Theta_{m}^{t} = fh(𝐞𝐭,𝐜,Θh)f_{h}(\mathbf{e^{t}},\mathbf{c},\Theta_{h}) = {fh(𝐞𝐭,𝐜1,Θh),fh(𝐞𝐭,𝐜2,Θh),,fh(𝐞𝐭,𝐜nc,Θh)}\{{f_{h}}(\mathbf{e^{t}},{\mathbf{c}}_{1},\Theta_{h}),{f_{h}}(\mathbf{e^{t}},{\mathbf{c}}_{2},\Theta_{h}),\ldots,{f_{h}}(\mathbf{e^{t}},{\mathbf{c}}_{n_{c}},\Theta_{h})\}, where ncn_{c} being the number of chunks. hypernetworks can still suffer from catastrophic forgetting when hypernetwork parameters are updated to generate main network parameters for the new task. In order to overcome this forgetting, an additional regularisation term is used while learning the parameters of the hypernetwork for a new task along with task specific loss [28].

Chunked hypernetworks generate weights of the main network in smaller chunks using a hypernetwork considering the task embedding and chunk embedding. We observe that they do not consider sequential nature and inter-dependence of weights between the chunks. We note that the chunked hypernetworks make conditional independence assumptions on the weights of the main network. Consequently, if we consider a probability distribution over weights, it gets decomposed over chunks as P(Θmt,1,Θmt,2,,Θmt,nc|𝐞𝐭,c)=P(Θmt,1|𝐞𝐭,𝐜1)×P(Θmt,2|𝐞𝐭,𝐜2)P(Θmt,nc|𝐞𝐭,𝐜nc)P({\Theta_{m}^{t,1}},{\Theta_{m}^{t,2}},{\ldots},{\Theta_{m}^{t,n_{c}}}|{\mathbf{e^{t}},\textbf{c}})=P({\Theta_{m}^{t,1}}|{\mathbf{e^{t}},\mathbf{c}_{1}})\times P({\Theta_{m}^{t,2}}|{\mathbf{e^{t}},\mathbf{c}_{2}})\ldots P({\Theta_{m}^{t,n_{c}}}|{\mathbf{e^{t}},\mathbf{c}_{n_{c}}}). Assumption of independence across the chunks does not usually hold and can affect the main network parameter generation and the performance of the continual learning.

3.1 LSTM Hypernetworks

To capture the interdependence in the chunk weights and to be parameter efficient at the same time, we propose a recurrent neural network (RNN) and in particular long short term memory (LSTM[9] based hypernetwork called LSTM_NET. LSTM_NET is capable of generating weights for main network in smaller chunks while also maintaining the dependencies across the chunks. LSTMs are sequence models capable of capturing long range dependencies and will be able to generate a chunk weight depending on the weights associated with the preceding chunks. Consequently, it models the joint probability over the main network parameters as

P(Θmt,1,Θmt,2,,Θmt,nc|𝐞𝐭,c)=P(Θmt,1|𝐞𝐭,𝐜1)×P(Θmt,2|Θmt,1,𝐞𝐭,𝐜2)P(Θmt,nc|Θmt,1Θmt,nc1,𝐞𝐭,𝐜nc)\begin{split}P({\Theta_{m}^{t,1}},{\Theta_{m}^{t,2}},{\ldots},{\Theta_{m}^{t,n_{c}}}|{\mathbf{e^{t}},\textbf{c}})&=P({\Theta_{m}^{t,1}}|{\mathbf{e^{t}},\mathbf{c}_{1}})\times\\ &\mkern-36.0mu\mkern-36.0mu\mkern-36.0mu\mkern-36.0mu\mkern-36.0mu\mkern-36.0muP({\Theta_{m}^{t,2}}|{\Theta_{m}^{t,1}},{\mathbf{e^{t}},\mathbf{c}_{2}})\ldots P({\Theta_{m}^{t,n_{c}}}|{\Theta_{m}^{t,1}}\ldots{\Theta_{m}^{t,{n_{c}-1}}},{\mathbf{e^{t}},\mathbf{c}_{n_{c}}})\end{split} (1)

Refer to caption

Figure 1: LSTM based Hypernetwork can generate main network weights Θmt\Theta_{m}^{t} for task tt in smaller chunks by using same hypernetwork iteratively but also by maintaining the dependencies across them using LSTM.

The proposed LSTM hypernetwork uses hidden state 𝐡𝐣𝟏𝐭\mathbf{h_{j-1}^{t}} and cell state 𝐬𝐣𝟏𝐭\mathbf{s_{j-1}^{t}} of preceding chunk along with current chunk embedding to generate the chunk weights. The following LSTM operations are used to generate the chunk weights assuming a single layer network where it takes task embedding 𝐞𝐭\mathbf{e^{t}} and chunk embedding 𝐜𝐣\mathbf{c_{j}} as input.

𝐢𝐣𝐭=σ(𝐰𝐢×𝐞𝐭,𝐜𝐣+𝐮𝐢×𝐡𝐣𝟏𝐭)\displaystyle\mathbf{i_{j}^{t}}=\sigma(\mathbf{w_{i}}\times{\langle\mathbf{e^{t}},\mathbf{c_{j}}\rangle}+\mathbf{u_{i}}\times\mathbf{h_{j-1}^{t}})
𝐟𝐣𝐭=σ(𝐰𝐟×𝐞𝐭,𝐜𝐣+𝐮𝐟×𝐡𝐣𝟏𝐭)\displaystyle\mathbf{f_{j}^{t}}=\sigma(\mathbf{w_{f}}\times{\langle\mathbf{e^{t}},\mathbf{c_{j}}\rangle}+\mathbf{u_{f}}\times\mathbf{h_{j-1}^{t}})
𝐨𝐣𝐭=σ(𝐰𝐨×𝐞𝐭,𝐜𝐣+𝐮𝐨×𝐡𝐣𝟏𝐭)\displaystyle\mathbf{o_{j}^{t}}=\sigma(\mathbf{w_{o}}\times{\langle\mathbf{e^{t}},\mathbf{c_{j}}\rangle}+\mathbf{u_{o}}\times\mathbf{h_{j-1}^{t}})
𝐠𝐣𝐭=Tanh(𝐰𝐠×𝐞𝐭,𝐜𝐣+𝐮𝐠×𝐡𝐣𝟏𝐭)\displaystyle\mathbf{g_{j}^{t}}=Tanh(\mathbf{w_{g}}\times{\langle\mathbf{e^{t}},\mathbf{c_{j}}\rangle}+\mathbf{u_{g}}\times\mathbf{h_{j-1}^{t}})
𝐬𝐣𝐭=𝐟𝐣𝐭𝐬𝐣𝟏𝐭+𝐢𝐣𝐭𝐠𝐣𝐭\displaystyle\mathbf{s_{j}^{t}}=\mathbf{f_{j}^{t}}\odot\mathbf{s_{j-1}^{t}}+\mathbf{i_{j}^{t}}\odot\mathbf{g_{j}^{t}}
𝐡𝐣𝐭=𝐨𝐣𝐭tanh(𝐬𝐣𝐭)\displaystyle\mathbf{h_{j}^{t}}=\mathbf{o_{j}^{t}}\odot\tanh(\mathbf{s_{j}^{t}})
Θm(t,j)=𝐡𝐣𝐭W\displaystyle\Theta_{m}^{(t,j)}=\mathbf{h_{j}^{t}}W

where 𝐟,𝐢,𝐨,𝐠\mathbf{f,i,o,g} are forget gate, input gate, output gate, and cell gate respectively. σ\sigma is the sigmoid function, and \odot is the Hadamard product. w, u are the weights associated with the input and hidden states respectively with subscript denoting the corresponding gate. Θm(t,j)\Theta_{m}^{(t,j)} is the jthj^{th} chunk of main network weights generated by LSTM hypernetwork. Here, WRd1×d2W\in R^{d_{1}\times d_{2}} are the weights of the feed forward layer producing the chunk weights.

We learn the LSTM parameters Θh\Theta_{h} when presented with data from task TT by minimising the following loss consisting of task specific loss and regularization loss [28].

argminΘhtotal=task(Θh,𝐞T,𝐜,XT,YT)+βT1t=1T1fh(𝐞t,𝐜,Θh)fh(𝐞t,𝐜,Θh+ΔΘh)2\begin{split}\operatorname*{arg\,min}_{\Theta_{h}}\mathcal{L}_{total}&=\mathcal{L}_{task}(\Theta_{h},\mathbf{e}^{T},\mathbf{c},X^{T},Y^{T})\\ &\mkern-36.0mu\mkern-36.0mu\mkern-18.0mu\mkern-18.0mu+\frac{\beta}{T-1}\sum_{t=1}^{T-1}{||}f_{h}(\mathbf{e}^{t},\mathbf{c},\Theta_{h}^{*})-f_{h}(\mathbf{e}^{t},\mathbf{c},\Theta_{h}+\Delta\Theta_{h}){||^{2}}\end{split} (2)

where fh(𝐞𝐭,𝐜,Θh)=Θmtf_{h}(\mathbf{e^{t}},\mathbf{c},\Theta_{h})=\Theta_{m}^{t} denotes all the parameters of the main network for task tt, Θh\Theta_{h}^{*} are parameters of the hypernetwork before learning task TT, β\beta is a regularisation constant which balances task specific loss and regularization loss, and ΔΘh\Delta\Theta_{h} is change in direction of the weights of the hypernetwork evaluated on the task-specific loss. The task-specific loss task\mathcal{L}_{task} is the loss associated with the task (for e.g., cross-entropy loss for classification). The regularization loss constraints the hypernetwork learnt on the new task to generate main network parameters similar to the ones generated by previously learnt hypernetwork. Hypernetwork parameters Θh\Theta_{h} and chunk embeddings are learnt by minimising the total loss total\mathcal{L}_{total}, and the task embeddings are learnt using task\mathcal{L}_{task} alone using backpropagation.

3.2 Importance Weighted Regularisation

We propose a novel regularisation technique for continual learning in hypernetworks which provides more flexibility to the hypernetwork to adapt to the new task compared to the regularisation in (2). The proposed importance weighted regularisation (IWR) updates the hypernetwork parameters based on the importance of the parameters associated with the main network for each task. IWR requires the hypernetwork to generate only important main network parameters associated with old tasks and not all the main network parameters. We achieve this by considering the fisher information score of the main network parameters in the regularisation term in (2). This will enforce hypernetwork to give importance to significant main network parameters during generation of main network weights. Meanwhile, it provides flexibility to the hypernetwork to adapt its parameters more freely to the new task as it is not constrained to generate all the main network parameters with equal importance. The objective function considering the IWR regularisation is defined as

argminΘhtotal=task(Θh,𝐞T,𝐜,XT,YT)+βT1t=1T1iFIit(fhi(𝐞t,𝐜,Θh)fhi(𝐞t,𝐜,Θh+ΔΘh))2\begin{split}\operatorname*{arg\,min}_{\Theta_{h}}\mathcal{L}_{total}&=\mathcal{L}_{task}(\Theta_{h},\mathbf{e}^{T},\mathbf{c},X^{T},Y^{T})\\ &\mkern-36.0mu\mkern-36.0mu\mkern-36.0mu\mkern-6.0mu+{\frac{\beta}{T-1}\sum_{t=1}^{T-1}\sum_{i}FI^{t}_{i}({{f_{h}}_{i}}(\mathbf{e}^{t},\mathbf{c},\Theta_{h}^{*})-{f_{h}}_{i}(\mathbf{e}^{t},\mathbf{c},\Theta_{h}+\Delta\Theta_{h}))^{2}}\end{split} (3)

where the first term task\mathcal{L}_{task} is task specific loss and the second term is the IWR term that regularises the hypernetwork parameters to avoid forgetting. The IWR term uses FIt{FI}^{t} which is the fisher information matrix (defined below) over the main network parameters associated with task tt. FIt{FI}^{t} provides the importance of the main network parameters for task tt, and the index ii iterates over all the main network parameters. fhi(𝐞t,𝐜,Θh){f_{h}}_{i}(\mathbf{e}^{t},\mathbf{c},\Theta_{h}) denotes the ithi^{th} main network parameter generated by the hypernetwork. We can observe that if the Fisher information associated with the ithi^{th} main network parameter is high (implying that this parameter is important), then hypernetwork is required to generate it exactly while it does not have to do the same for the unimportant parameters. Thus, IWR provides more flexibility to the hypernetwork to learn and adapt to the new tasks.

The Fisher information matrix (FI) provides the information on the importance of each weight in the network. For the IWR in (3), the FI matrix is defined as

FIt=1Nti=1Nt[ΘmtLtask(Θh,𝐞t,yi,𝐱i)×\displaystyle FI^{t}=\frac{1}{N_{t}}\sum_{i=1}^{N_{t}}\biggl{[}\nabla_{\Theta_{m}^{t}}L_{task}(\Theta_{h},\mathbf{e}^{t},y_{i},\mathbf{x}_{i})\times
ΘmtLtask(Θh,𝐞t,yi,𝐱i)T]\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\nabla_{\Theta_{m}^{t}}L_{task}(\Theta_{h},\mathbf{e}^{t},y_{i},\mathbf{x}_{i})^{T}\biggl{]} (4)

We note that the derivatives are computed with respect to the main network parameters to assess the importance of those parameters and not with respect to the hypernetwork parameters that we are learning unlike the standard regularization techniques. Thus, using Fisher information matrix we can find out the main network parameters which are important in learning the task. The existing regularisation for hypernetworks in (2) treats all the main network parameters equally. In practice, not all main network parameters contribute equally to solving a particular task. Hence, it is not required to exactly generate all the main network parameters by the hypernetwork but only important ones and can be achieved using the IWR regularisation. The IWR regularization is a generic technique that can be used with any hypernetwork and not only the LSTM hypernetwork to improve the continual learning performance.

3.3 Network Growth Technique

One potential problem with the regularisation approaches is that training time grows with number of tasks as can be seen from Eq (2) and Eq (3). Moreover, the same hypernetwork parameters are used to generate all the task specific main parameters. This can become a bottleneck and affects continual learning performance in situations with a large number of tasks. We propose LSTM hypernetworks (LSTM_NET_GROW) based on network growth for continual learning. It provides more flexibility in adapting to the new task by maintaining task specific parameters and accelerates model training by requiring no regularisation. In order to transfer knowledge across the tasks, LSTM_NET_GROW also maintains a shared set of hypernetwork parameters.

We hypothesize that, though the actual main network parameters differ across the tasks, the dependencies existing among the main network parameters remain the same across the tasks. Based on this intuition, we define the shared and task specific parameters in the LSTM hypernetwork. In LSTM, the dependencies are captured by the weights associated with the hidden state. Hence, we assume them to be the same across the tasks in the proposed LSTM_NET_GROW model. The variability in parameter generation across the tasks is captured by having task specific weights associated with inputs. More specifically, the weights (𝐮𝐟,𝐮𝐢,𝐮𝐨,𝐮𝐠\mathbf{u_{f},u_{i},u_{o},u_{g}}) of the LSTM are shared across the tasks and we maintain input weights (𝐰𝐟𝐭,𝐰𝐢𝐭,𝐰𝐨𝐭,𝐰𝐠𝐭\mathbf{w_{f}^{t},w_{i}^{t},w_{o}^{t},w_{g}^{t}}) to be task specific. LSTM_NET_GROW uses the following LSTM operations to generate chunk weights of the main network parameters associated with task tt.

𝐢𝐣𝐭=σ(𝐰𝐢𝐭×𝐞𝐭,𝐜𝐣+𝐮𝐢×𝐡𝐣𝟏𝐭)\displaystyle\mathbf{i_{j}^{t}}=\sigma(\mathbf{w_{i}^{t}}\times{\langle\mathbf{e^{t}},\mathbf{c_{j}}\rangle}+\mathbf{u_{i}}\times\mathbf{h_{j-1}^{t}})
𝐟𝐣𝐭=σ(𝐰𝐟𝐭×𝐞𝐭,𝐜𝐣+𝐮𝐟×𝐡𝐣𝟏𝐭)\displaystyle\mathbf{f_{j}^{t}}=\sigma(\mathbf{w_{f}^{t}}\times{\langle\mathbf{e^{t}},\mathbf{c_{j}}\rangle}+\mathbf{u_{f}}\times\mathbf{h_{j-1}^{t}})
𝐨𝐣𝐭=σ(𝐰𝐨𝐭×𝐞𝐭,𝐜𝐣+𝐮𝐨×𝐡𝐣𝟏𝐭)\displaystyle\mathbf{o_{j}^{t}}=\sigma(\mathbf{w_{o}^{t}}\times{\langle\mathbf{e^{t}},\mathbf{c_{j}}\rangle}+\mathbf{u_{o}}\times\mathbf{h_{j-1}^{t}})
𝐠𝐣𝐭=Tanh(𝐰𝐠𝐭×𝐞𝐭,𝐜𝐣+𝐮𝐠×𝐡𝐣𝟏𝐭)\displaystyle\mathbf{g_{j}^{t}}=Tanh(\mathbf{w_{g}^{t}}\times{\langle\mathbf{e^{t}},\mathbf{c_{j}}\rangle}+\mathbf{u_{g}}\times\mathbf{h_{j-1}^{t}})
𝐬𝐣𝐭=𝐟𝐣𝐭𝐬𝐣𝟏𝐭+𝐢𝐣𝐭𝐠𝐣𝐭\displaystyle\mathbf{s_{j}^{t}}=\mathbf{f_{j}^{t}}\odot\mathbf{s_{j-1}^{t}}+\mathbf{i_{j}^{t}}\odot\mathbf{g_{j}^{t}}
𝐡𝐣𝐭=𝐨𝐣𝐭tanh(𝐬𝐣𝐭)\displaystyle\mathbf{h_{j}^{t}}=\mathbf{o_{j}^{t}}\odot\tanh(\mathbf{s_{j}^{t}})
𝚯𝐦(𝐭,𝐣)=𝐡𝐣𝐭𝐖𝐭\displaystyle\mathbf{\Theta_{m}^{(t,j)}}=\mathbf{h_{j}^{t}W^{t}}
Table 1: Comparison of average test accuracy (%) of Split MNIST and Permuted MNIST for all three scenarios of continual learning without generative replay
Split MNIST Permuted MNIST
CL1 CL2 CL3 CL1 CL2 CL3
EWC [11] 98.64±\pm0.22 63.95±\pm1.90 20.10±\pm0.06 94.74±\pm0.05 94.31±\pm0.11 25.04±\pm0.50
Online EWC[11] 99.12±\pm0.11 64.32±\pm1.90 19.96±\pm0.07 95.96±\pm0.06 94.42±\pm0.13 33.88±\pm0.49
SI[30] 99.09±\pm0.15 65.36±\pm1.57 19.99±\pm0.06 94.75±\pm0.14 95.33±\pm0.11 29.31±\pm0.62
HNET[28] 99.79±\pm0.01 87.01±\pm0.47 69.48±\pm0.80 97.57±\pm0.02 92.80±\pm0.15 91.75±\pm0.21
HNET_IWR 99.79±\pm0.01 88.51±\pm0.18 71.90±\pm0.11 97.60±\pm0.04 93.90±\pm0.11 92.15±\pm0.19
LSTM_NET 99.82±\pm0.01 89.50±\pm0.19 71.31±\pm0.07 97.65±\pm0.01 93.11±\pm0.13 92.10±\pm0.20
LSTM_NET_IWR 99.85±\pm0.02 90.17±\pm0.25 71.54±\pm0.04 97.74±\pm0.03 94.26±\pm0.10 92.21±\pm0.23
LSTM_NET_GROW 99.85±\pm0.02 97.11±\pm0.16 83.21±\pm0.02 97.88±\pm0.02 95.46±\pm0.09 92.23±\pm0.19
Table 2: Comparison of average test accuracy (%) of Split MNIST and Permuted MNIST for all three scenarios of continual learning with generative replay
Split MNIST Permuted MNIST
CL1 CL2 CL3 CL1 CL2 CL3
LWF[14] 99.57±\pm0.02 71.50±\pm1.63 23.85±\pm0.44 69.84±\pm0.46 72.64±\pm0.52 22.64±\pm0.23
DGR[25] 99.50±\pm0.03 95.72±\pm0.25 90.79±\pm0.41 92.52±\pm0.08 95.09±\pm0.04 92.19±\pm0.09
DGR+distill[25] 99.61±\pm0.02 96.83±\pm0.20 91.79±\pm0.32 97.51±\pm0.01 97.35±\pm0.02 96.38±\pm0.03
HNET+R[28] 99.83±\pm0.01 98.00±\pm0.03 95.30±\pm0.13 97.87±\pm0.01 97.60±\pm0.01 97.76±\pm0.01
HNET_IWR+R 99.83±\pm0.01 97.94±\pm0.05 95.38±\pm0.16 97.85±\pm0.01 97.66±\pm0.02 97.76±\pm0.02
LSTM_NET+R 99.83±\pm0.01 98.17±\pm0.02 95.46±\pm0.11 97.87±\pm0.01 97.60±\pm0.01 97.77±\pm0.01
LSTM_NET_IWR+R 99.83±\pm0.01 98.39±\pm0.05 96.50±\pm0.19 97.87±\pm0.01 97.66±\pm0.01 97.80±\pm0.02
LSTM_NET_GROW+R 99.83±\pm0.01 98.43±\pm0.05 97.01±\pm0.13 97.90±\pm0.01 97.70±\pm0.02 97.80±\pm0.01

The LSTM_NET_GROW model freezes the hidden weights (𝐮𝐟,𝐮𝐢,𝐮𝐨,𝐮𝐠\mathbf{u_{f},u_{i},u_{o},u_{g}}) of the LSTM after learning first task and is shared across all the tasks. It keeps learning new task specific input weights (𝐰𝐟𝐭,𝐰𝐢𝐭,𝐰𝐨𝐭,𝐰𝐠𝐭\mathbf{w_{f}^{t},w_{i}^{t},w_{o}^{t},w_{g}^{t}}) upon training on the new task and those are stored for inference at a later stage. This approach does not require the additional regularisation term and can be learnt just based on task specific loss (LtaskL_{task}). In addition, the task specific weights provide extra flexibility to the LSTM hypernetwork in generating task specific main network weights.

4 Experiments

We perform extensive experiments on various continual learning setups and real world benchmarks datasets to show the effectiveness of our approach. We present our results on Split MNIST, permuted MNIST, CIFAR-10, and CIFAR-100 datasets. Through experiments we aim to demonstrate:

  • impact of maintaining dependencies across the chunks using the LSTM based hypernetwork (LSTM_NET),

  • impact of the proposed regularisation IWR on LSTM_NET (LSTM_NET_IWR) and on HNET (HNET_IWR) in mitigating catastrophic forgetting,

  • improvement in performance using proposed dynamically growing LSTM based hypernetwork LSTM_NET_GROW.

  • knowledge transfer and mitigating forgetting across the tasks using the challenging Cifar datasets.

4.1 Experimental Setup

Continual learning models are tested on three different continual learning scenarios [27].
CL1 (Task incremental learning): It provides the task identity information to the model both at training and testing time. Since task identity is available, dedicated components can be assigned to the tasks in the sequence. Multi-headed model is one such kind of model architecture used in continual learning.
CL2 (Domain incremental learning): It does not provide task identity information at test time, nor is it required to infer the task identity. Here, each task treats data to be from different domains but with the same classes. It considers the same number of classes across all the tasks in the sequence and utilises the same output heads for all the tasks.
CL3 (Class incremental learning): In this continual learning scenario, the model is not provided with the task identity. It not only requires learning the tasks incrementally but also requires to infer the task identity. Here, the task identity is inferred through the predictive distribution entropy. This scenario resembles most with the real time setting with new class objects appearing incrementally.

Table 3: Test accuracy comparison of various methods on CIFAR-10 (C-10) and subsequent five splits (S15S_{1\ldots 5} ) each with ten classes of CIFAR-100 (C-100).
C-10 (%) C-100 S1S_{1}(%) C-100 S2S_{2}(%) C-100 S3S_{3}(%) C-100 S4S_{4}(%) C-100 S5S_{5}(%) Average-accuracy(%)
Finetuning 15.3 13.1 12.2 10.2 20.3 87.0 26.35
Training-from-scratch 88.6 79.3 77.0 83.0 74.4 81.1 80.60
HNET[28] 88 83 79 82 81 82 82.50
HNET_IWR 86.2 86.1 80.2 85.0 83.6 85.01 84.35
LSTM_NET_during 88.78 89.3 85.2 84.4 83.5 82.7 85.64
LSTM_NET 88.74 89.1 84.9 84.3 83.4 82.7 85.52
LSTM_NET_IWR 88.44 88.9 85.2 88.5 86.3 86.8 87.35
LSTM_NET_GROW 88.98 87.7 86.3 88.2 89.2 89.1 88.25

The effectiveness of the hypernetwork based CL techniques for parameter generation are also tested on two continual learning setups, replay based and non-replay based setups. In the non-replay-based setup, hypernetwork is trained to generate parameters of the classifier used for solving the image classification problems. In the non-replay based setup, we compare the proposed approach with regularisation baselines Elastic Weight Consolidation (EWC) [11], Synaptic Intelligence (SI) [30] and the baseline hypernetwork HNET [28]. In the replay based setup, we augment our system with a generative model, for e.g. variational auto-encoder (VAE) to generate synthetic examples from the previous tasks that can be replayed to aid the classifier to remember previous tasks. In this case, hypernetwork will generate weights for the replay network i.e. VAE but not the target classifier. In the replay based setup, we compare the proposed approach with baselines deep generative replay with distillation (DGR)[25], learning without forgetting (LWF)[14] and the baseline hypernetwork HNET[28].

We conduct experiments on the standard continual learning task of image classification on publicly available real world data sets such as split MNIST, permuted MNIST, CIFAR-10 and CIFAR-100. In these experiments, we use a single layer LSTM that takes a task and chunk embedding each of size 96. For MNIST, we consider an LSTM with hidden state size 64 and batch size 128. For CIFAR, hidden size and batch size are 128 and 32 respectively. For Split MNIST, the classifier is a fully connected network (FCN) with 2 layers each with size 400 [28]. For Permuted MNIST, layer size is taken to be 1000 as done in [28]. For CIFAR datasets, Resnet-32 is used as the classifier. In replay based setup, we use VAE which uses an FCN with two layers each of size 400 both as the encoder and the decoder and uses a latent space of dimension 100.

4.2 Results

We demonstrate the results of our experiments on various image classification datasets used for the continual learning setup and several baselines. For fair comparison, we maintained the number of trainable parameters in the hypernetwork to be equal or lesser than the baseline methods.

4.2.1 Split MNIST:

Split MNIST is a popular continual learning benchmark for image classification. The dataset consists of images of ten digits (0-9) and form five binary classification tasks by pairing them sequentially i.e.{(0,1),(2,3),(4,5),\{(0,1),(2,3),(4,5), (6,7),(8,9)}(6,7),(8,9)\}. The results are presented in Table 1 for non-replay based setup, and in Table 2 for the replay based setup. The results show the efficacy of our approach in achieving better continual learning performance in all the three CL scenarios and for each of the setups.

The proposed hypernetwork LSTM_NET outperforms the baselines EWC, SI and HNET in the non-replay based setup, and outperforms the baselines LWF, DGR and HNET in the replay based setup. The performance of HNET_IWR and LSTM_NET_IWR demonstrates that the proposed regularisation technique IWR further improves the performance of HNET and LSTM_NET in all the continual learning setups. While the methods provide comparable results for the easier CL1 setting, the improvement in performance using the proposed techniques are more evident in the more complex and realistic CL2 and CL3 settings. This is even more evident in the non-replay setup where the standard techniques struggle. We can observe that the proposed approach LSTM_NET_GROW has significantly improved the continual learning performance over the other models for these CL scenarios and setups. One of the major reasons for large improvement in accuracy of CL2 and CL3 with LSTM_NET_GROW is because of dynamically expanding network with new tasks, this helps each task to have task-specific parameters which won’t get updated while learning new tasks and contribute significantly for improvement in performance.

4.2.2 Permuted MNIST

This CL benchmark is a variant of MNIST, it consists of tasks which are created by performing random permutation over MNIST images. The sequence of T=10T=10 tasks are obtained by repeating this procedure. We consider a dataset with a sufficiently long sequence of tasks (T=10T=10) to investigate the remembering capacity of our continual learning model. The results presented in Table 1 and Table 2 for non-replay and replay based setups respectively demonstrates the effectiveness of the proposed approaches for continual learning on Permuted MNIST. The results follows a similar trend as in the split MNIST, with LSTM_NET_GROW giving the best results, followed by LSTM_NET_IWR and LSTM_NET, beating the baseline approaches.

4.2.3 CIFAR-10/100 dataset

We further evaluate the effectiveness of the proposed approaches on a more challenging image classification data CIFAR-10 and CIFAR-100. The model is first trained on 10 classes of CIFAR-10, and subsequently on five sets of ten classes from cifar-100 following the experimental setup in [28]. Thus, the model needs to learn T=6T=6 tasks. We use ResNet-32 for classification of CIFAR-10/100 datasets and the hypernetworks are trained to generate parameters of ResNet-32 architecture. The experiments are conducted on the CL1 scenario and non-replay based setup following [28]. In addition to the baseline hypernetwork HNET, we also consider baselines which allows us to demonstrate the knowledge transfer across the tasks. Training-from-scratch baseline independently and separately learns the main network parameters for each task and test the performance on the corresponding task. The baseline Finetuning adapts the main network to the new tasks without taking into account catastrophic forgetting. The model adapted to the final task is then used for predicting performance on all the tasks. To demonstrate effectiveness of LSTM_NET in dealing with catastrophic forgetting, we also considered a baseline LSTM_NET_during, where we test LSTM_NET on each task immediately after training on that task instead of testing after training on all the tasks as in LSTM_NET.

We provide the results comparing all teh approaches in Table 3. We can clearly observe from the results that our approach LSTM_NET outperforms HNET by a great margin on a challenging dataset like CIFAR-10/100. In Table 3, comparing results of LSTM_NET, LSTM_NET_IWR and LSTM_NET_GROW with Training-from-scratch, we can see that the knowledge transfer across the tasks helps the proposed approaches in getting a better performance. We can also observe that the LSTM_NET_during matches with LSTM_NET which indicates that LSTM based hypernetwork is very effective in dealing with catastrophic forgetting. We also perform experiments using different regularisation techniques. The proposed IWR regularisation achieves a better result than the baseline regularisation proposed in [28]. In fact, it improves the HNET performance as well, an improvement in the overall test accuracy by almost 2%. Hence, IWR is an effective regularisation technique for any hypernetwork based continual learning approaches. The performance improves further by using the LSTM_NET_GROW approach for continual learning in this data similar to MNIST.

4.2.4 Ablation Study

We conduct further ablation study on CIFAR-10/CIFAR-100 to understand the impact of compression ratio (Figure LABEL:fig:3.1) and regularisation constant (Figure LABEL:fig:3.2) on the proposed models. From Figure LABEL:fig:3.1, we can see that as the number of trainable parameters grows in hypernetwork compared to the main network, LSTM_NET performance further improves over the HNET. In Figure LABEL:fig:3.2, we study the effect of varying the regularisation constant (β\beta) in the IWR regularization term in LSTM_NET_IWR on Cifar datasets. The performance is poor when regularization term is neglected (low value of β\beta) as expected and is high and stable for higher values of β\beta.

From the experiments on CIFAR-10/CIFAR-100 tasks, we observe that training time using LSTM_NET with regularisations in Eq.2 and Eq.3 is approximately 28 hours but with LSTM_NET_GROW it is approximately 20 hours. The above results are calculated with a batch size of 32 and 200 epochs for each batch. As the number of tasks increases, training time of regularisation approaches grows linearly but that of LSTM_NET_GROW remains constant. On the other hand, the memory requirements of the regularisation approaches remains constant but LSTM_NET_GROW grows linearly with tasks due to some task specific parameters but is much lower than maintaining separate parameters for each task.

Refer to caption
(a)
Refer to caption
(b)
Figure 2: In Figure LABEL:fig:3.1, plotted results of accuracies of LSTM_NET vs HNET with increase in compression ratio (ratio of trainable parameters in hypernetwork to main network) on Cifar datasets. In Figure LABEL:fig:3.2, plotted graph with accuracy values using LSTM_NET_IWR with varying regularisation constant (β\beta) in the IWR on Cifar datasets.

5 Conclusion

We propose a novel LSTM based hypernetwork for continual learning which could capture dependencies across the main network parameters while also maintaining the parameter efficiency. To improve continual learning performance using hypernetworks, we propose a novel regularisation Importance Weighted Regularisation (IWR) which is well-suited for hypernetwork based CL approaches. To further improve the continual learning performance of the proposed LSTM hypernetwork, we propose a network growth technique for LSTMs. Through experiments on several image classification tasks and datasets, we demonstrate the effectiveness of our proposed approaches, LSTMLSTM based hypernetwork, IWR regularisation for hypernetworks, and network growth on LSTMs. The proposed approaches improved the continual learning performance on all the CL tasks, settings, and datasets. As a future work, we would like to improve the parameter growth in LSTM_NET_GROW, and develop hybrid models combining the network growth and regularisation to further improve CL performance.

References

  • [1] Rahaf Aljundi, Francesca Babiloni, Mohamed Elhoseiny, Marcus Rohrbach, and Tinne Tuytelaars. Memory aware synapses: Learning what (not) to forget. In Vittorio Ferrari, Martial Hebert, Cristian Sminchisescu, and Yair Weiss, editors, Computer Vision - ECCV 2018 - 15th European Conference, Munich, Germany, September 8-14, 2018, Proceedings, Part III, volume 11207 of Lecture Notes in Computer Science, pages 144–161. Springer, 2018.
  • [2] Arslan Chaudhry, Marc’Aurelio Ranzato, Marcus Rohrbach, and Mohamed Elhoseiny. Efficient lifelong learning with A-GEM. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net, 2019.
  • [3] Arslan Chaudhry, Marcus Rohrbach, Mohamed Elhoseiny, Thalaiyasingam Ajanthan, Puneet K Dokania, Philip HS Torr, and Marc’Aurelio Ranzato. On tiny episodic memories in continual learning. arXiv preprint arXiv:1902.10486, 2019.
  • [4] Benjamin Ehret, Christian Henning, Maria R. Cervera, Alexander Meulemans, Johannes von Oswald, and Benjamin F. Grewe. Continual learning in recurrent neural networks with hypernetworks. CoRR, abs/2006.12109, 2020.
  • [5] Sibo Gai, Zhengyu Chen, and Donglin Wang. Multi-modal meta continual learning. In International Joint Conference on Neural Networks, IJCNN 2021, Shenzhen, China, July 18-22, 2021, pages 1–8. IEEE, 2021.
  • [6] Chandan Gautam, Sethupathy Parameswaran, Ashish Mishra, and Suresh Sundaram. Generalized continual zero-shot learning. CoRR, abs/2011.08508, 2020.
  • [7] Stephen Grossberg. Consciousness CLEARS the mind. Neural Networks, 20(9):1040–1053, 2007.
  • [8] David Ha, Andrew M. Dai, and Quoc V. Le. Hypernetworks. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. OpenReview.net, 2017.
  • [9] Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural computation, 9(8):1735–1780, 1997.
  • [10] Ronald Kemker, Marc McClure, Angelina Abitino, Tyler L. Hayes, and Christopher Kanan. Measuring catastrophic forgetting in neural networks. In Sheila A. McIlraith and Kilian Q. Weinberger, editors, Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, (AAAI-18), the 30th innovative Applications of Artificial Intelligence (IAAI-18), and the 8th AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI-18), New Orleans, Louisiana, USA, February 2-7, 2018, pages 3390–3398. AAAI Press, 2018.
  • [11] James Kirkpatrick, Razvan Pascanu, Neil Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei A Rusu, Kieran Milan, John Quan, Tiago Ramalho, Agnieszka Grabska-Barwinska, et al. Overcoming catastrophic forgetting in neural networks. Proceedings of the national academy of sciences, 114(13):3521–3526, 2017.
  • [12] Jan Koutník, Faustino J. Gomez, and Jürgen Schmidhuber. Evolving neural networks in compressed weight space. In Martin Pelikan and Jürgen Branke, editors, Genetic and Evolutionary Computation Conference, GECCO 2010, Proceedings, Portland, Oregon, USA, July 7-11, 2010, pages 619–626. ACM, 2010.
  • [13] Matthias De Lange, Rahaf Aljundi, Marc Masana, Sarah Parisot, Xu Jia, Ales Leonardis, Gregory G. Slabaugh, and Tinne Tuytelaars. Continual learning: A comparative study on how to defy forgetting in classification tasks. CoRR, abs/1909.08383, 2019.
  • [14] Zhizhong Li and Derek Hoiem. Learning without forgetting. In Bastian Leibe, Jiri Matas, Nicu Sebe, and Max Welling, editors, Computer Vision - ECCV 2016 - 14th European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part IV, volume 9908 of Lecture Notes in Computer Science, pages 614–629. Springer, 2016.
  • [15] Xialei Liu, Marc Masana, Luis Herranz, Joost van de Weijer, Antonio M. López, and Andrew D. Bagdanov. Rotate your networks: Better weight consolidation and less catastrophic forgetting. In 24th International Conference on Pattern Recognition, ICPR 2018, Beijing, China, August 20-24, 2018, pages 2262–2268. IEEE Computer Society, 2018.
  • [16] Noel Loo, Siddharth Swaroop, and Richard E. Turner. Generalized variational continual learning. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. OpenReview.net, 2021.
  • [17] David Lopez-Paz and Marc’Aurelio Ranzato. Gradient episodic memory for continual learning. In Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan, and Roman Garnett, editors, Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, pages 6467–6476, 2017.
  • [18] Michael McCloskey and Neal J Cohen. Catastrophic interference in connectionist networks: The sequential learning problem. In Psychology of learning and motivation, volume 24, pages 109–165. Elsevier, 1989.
  • [19] Martial Mermillod, Aurélia Bugaiska, and Patrick Bonin. The stability-plasticity dilemma: Investigating the continuum from catastrophic forgetting to age-limited learning effects, 2013.
  • [20] German Ignacio Parisi, Ronald Kemker, Jose L. Part, Christopher Kanan, and Stefan Wermter. Continual lifelong learning with neural networks: A review. Neural Networks, 113:54–71, 2019.
  • [21] Sylvestre-Alvise Rebuffi, Alexander Kolesnikov, Georg Sperl, and Christoph H. Lampert. icarl: Incremental classifier and representation learning. In 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017, pages 5533–5542. IEEE Computer Society, 2017.
  • [22] Matthew Riemer, Ignacio Cases, Robert Ajemian, Miao Liu, Irina Rish, Yuhai Tu, and Gerald Tesauro. Learning to learn without forgetting by maximizing transfer and minimizing interference. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net, 2019.
  • [23] Andrei A. Rusu, Neil C. Rabinowitz, Guillaume Desjardins, Hubert Soyer, James Kirkpatrick, Koray Kavukcuoglu, Razvan Pascanu, and Raia Hadsell. Progressive neural networks. CoRR, abs/1606.04671, 2016.
  • [24] Jürgen Schmidhuber. Learning to control fast-weight memories: An alternative to dynamic recurrent networks. Neural Comput., 4(1):131–139, 1992.
  • [25] Hanul Shin, Jung Kwon Lee, Jaehong Kim, and Jiwon Kim. Continual learning with deep generative replay. In Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan, and Roman Garnett, editors, Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, pages 2990–2999, 2017.
  • [26] Kenneth O. Stanley, David B. D’Ambrosio, and Jason Gauci. A hypercube-based encoding for evolving large-scale neural networks. Artif. Life, 15(2):185–212, 2009.
  • [27] Gido M. van de Ven and Andreas S. Tolias. Three scenarios for continual learning. CoRR, abs/1904.07734, 2019.
  • [28] Johannes von Oswald, Christian Henning, João Sacramento, and Benjamin F. Grewe. Continual learning with hypernetworks. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net, 2020.
  • [29] Jaehong Yoon, Eunho Yang, Jeongtae Lee, and Sung Ju Hwang. Lifelong learning with dynamically expandable networks. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings. OpenReview.net, 2018.
  • [30] Friedemann Zenke, Ben Poole, and Surya Ganguli. Continual learning through synaptic intelligence. In Doina Precup and Yee Whye Teh, editors, Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017, volume 70 of Proceedings of Machine Learning Research, pages 3987–3995. PMLR, 2017.