This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

On Lyapunov Exponents for RNNs:
Understanding Information Propagation Using Dynamical Systems Tools

Ryan Vogt
Department of Applied Mathematics
University of Washington
ravogt95@uw.edu &Maximilian Puelma Touzel
Math & Stats Dept., Université de Montréal
Mila, Québec AI Institute
puelmatm@mila.quebec &Eli Shlizerman∗∗
Depts. of App. Math. and Elec. &\&
Comp. Engineering, University of Washington
shlizee@uw.edu &Guillaume Lajoie∗∗
Math & Stats Dept., Université de Montréal
Mila, Québec AI Institute
g.lajoie@umontreal.ca
Abstract

Recurrent neural networks (RNNs) have been successfully applied to a variety of problems involving sequential data, but their optimization is sensitive to parameter initialization, architecture, and optimizer hyperparameters. Considering RNNs as dynamical systems, a natural way to capture stability, i.e., the growth and decay over long iterates, are the Lyapunov Exponents (LEs), which form the Lyapunov spectrum. The LEs have a bearing on stability of RNN training dynamics because forward propagation of information is related to the backward propagation of error gradients. LEs measure the asymptotic rates of expansion and contraction of nonlinear system trajectories, and generalize stability analysis to the time-varying attractors structuring the non-autonomous dynamics of data-driven RNNs. As a tool to understand and exploit stability of training dynamics, the Lyapunov spectrum fills an existing gap between prescriptive mathematical approaches of limited scope and computationally-expensive empirical approaches. To leverage this tool, we implement an efficient way to compute LEs for RNNs during training, discuss the aspects specific to standard RNN architectures driven by typical sequential datasets, and show that the Lyapunov spectrum can serve as a robust readout of training stability across hyperparameters. With this exposition-oriented contribution, we hope to draw attention to this under-studied, but theoretically grounded tool for understanding training stability in RNNs.

*: authors share first authorship, order selected at random.

**: authors share senior authorship, order selected at random.

1 Introduction

The propagation of error gradients in deep learning leads to the study of recursive compositions and their stability [1]. Vanishing and exploding gradients arise from long products of Jacobians of the hidden state dynamics whose norm exponentially grows or decays, hindering training [2]. To mitigate this sensitivity, much effort has been made to mathematically understand the link between model parameters and the the eigen- and singular-value spectra of these long products [3, 4, 5]. For architectures used in practice, this approach appears to have limited a scope [6, 7] likely due to spectra having non-trivial properties reflecting complicated long-time dependencies within the trajectory. Fortunately, the theory of dynamical systems has formulated this problem for general dynamical systems, i.e., for arbitrary architectures. The Lyapunov spectrum (LS) formed by the Lyapunov exponents (LEs) is precisely the measurement that characterizes rates of expansion and contraction of nonlinear system dynamics.

While there are results using LEs in machine learning [8, 9, 10], most only look at the largest exponent, the sign of which indicates the presence or absence of chaos. Additionally, to our knowledge, current machine learning uses of LEs are for the autonomous case, i.e. , analyzing RNN dynamics in the absence of inputs, a regime at odds with the most common use of RNNs as input-processing systems. Thus, there is likely much to be gained from both the study of other features of the LS and from including input-driven regimes into LEs analysis. Indeed, a concurrent work coming from theoretical neuroscience takes up this approach [11] and outlines the ongoing work using LEs to study models of neural networks in the brain. However, the machinery to compute the LS and the theory surrounding its interpretation is still relatively unknown in the machine learning community.

In this exposition-style paper, we aim to fill this gap by presenting an overview of LEs in the context of RNNs, and discuss their usefulness for studying training dynamics. We present encouraging results which suggest that LEs could serve in efficient hyperparameter tuning by allowing early detection of performance. To compute LEs, we present a novel algorithm for various RNN architectures, taking advantage of modern deep learning environments. Finally, we highlight key directions of research that have the potential to leverage LEs for improvement of RNNs robustness and performance.

2 Motivation and Definitions

Here we develop the problem of spectral constraints for robust gradient propagation. For transparent exposition, in this section we will consider the "vanilla" RNN defined as

𝐨t=𝐖𝐡t,𝐡t\displaystyle\mathbf{o}_{t}=\mathbf{W}\mathbf{h}_{t},\;\mathbf{h}_{t} =ϕ(𝐚t),𝐚t=𝐕𝐡t1+𝐔𝐱t+𝐛,\displaystyle=\phi(\mathbf{a}_{t}),\;\mathbf{a}_{t}=\mathbf{V}\mathbf{h}_{t-1}+\mathbf{U}\mathbf{x}_{t}+\mathbf{b}\;, (1)

where 𝐕\mathbf{V} is the recurrent weight matrix that couples elements of the hidden state vector 𝐡tN\mathbf{h}_{t}\in\mathbb{R}^{N}, 𝐔\mathbf{U} projects the input, 𝐱t\mathbf{x}_{t}, into the network, 𝐛\mathbf{b} is a constant bias vector, and ϕ\phi applies a nonlinear scalar transformation element-wise. The readout weights, 𝐖\mathbf{W}, output the activity into the output variable, 𝐨t\mathbf{o}_{t}. The loss over TT iterates is L=t=1TLtL=\sum_{t=1}^{T}L_{t}, with Lt=f(𝐲t,𝐲^t)L_{t}=f(\mathbf{y}_{t},\hat{\mathbf{y}}_{t}), with ff some scalar loss function (e.g., cross-entropy loss, f(𝐲,𝐲^)=𝐲log𝐲^f(\mathbf{y},\hat{\mathbf{y}})={\mathbf{y}}\cdot\log\hat{\mathbf{y}}), 𝐲^t\hat{\mathbf{y}}_{t} the prediction (e.g., 𝐲t^=softmax(𝐨t)\hat{\mathbf{y}_{t}}=\mathrm{softmax}(\mathbf{o}_{t})), and 𝐲t\mathbf{y}_{t} is a one-hot binary target vector. The parameters, Θ=(𝐕,𝐔,𝐛,𝐖)\Theta=(\mathbf{V},\mathbf{U},\mathbf{b},\mathbf{W}), are learned by following the gradient of the loss, e.g. in the space of recurrent weights 𝐕\mathbf{V},

𝐕L=t=1Ti=1NLht,i𝐕ht,i=t=1Tdiag(ϕ(𝐚t))𝐡tL𝐡t1,\displaystyle\nabla_{\mathbf{V}}L=\sum_{t=1}^{T}\sum_{i=1}^{N}\frac{\partial L}{\partial h_{t,i}}\nabla_{\mathbf{V}}h_{t,i}=\sum_{t=1}^{T}\textrm{diag}(\phi^{\prime}(\mathbf{a}_{t}))\nabla_{\mathbf{h}_{t}}L\;\mathbf{h}_{t-1}^{\top}\;, (2)

where diag(𝐱)\textrm{diag}(\mathbf{x}) is the diagonal matrix formed by the vector 𝐱\mathbf{x} and ϕ\phi^{\prime} is the derivative of ϕ\phi and denotes transpose. Here,

𝐡tL=s=tT(r=t+1s𝐉r)𝐖𝐨sL,\displaystyle\nabla_{\mathbf{h}_{t}}L=\sum_{s=t}^{T}\left(\prod_{r=t+1}^{s}\mathbf{J}^{\top}_{r}\right)\mathbf{W}^{\top}\nabla_{\mathbf{o}_{s}}L\;, (3)

where 𝐨sL\nabla_{\mathbf{o}_{s}}L is some simple expression (e.g. 𝐲^𝐲t\hat{\mathbf{y}}-\mathbf{y}_{t} for cross-entropy loss) and 𝐉t=𝐡t𝐡t1\mathbf{J}_{t}=\frac{\partial\mathbf{h}_{t}}{\partial\mathbf{h}_{t-1}} is the Jacobian of the hidden state dynamics,

𝐉t\displaystyle\mathbf{J}_{t} =diag(ϕ(𝐚t))𝐕.\displaystyle=\textrm{diag}\left(\phi^{\prime}(\mathbf{a}_{t})\right)\mathbf{V}\;. (4)

𝐉t\mathbf{J}_{t} varies in time with 𝐱t\mathbf{x}_{t} and 𝐡t1\mathbf{h}_{t-1} via 𝐚t\mathbf{a}_{t} and so is treated as a random matrix with ensemble properties arising from the specified input statistics and the emergent hidden state statistics.

Gradients can vanish or explode with TT according to the singular value spectra of the products of Jacobians appearing in Eq. (3[1]. This fact motivated the study of hyperparameter constraints that control the average of the latter over the input distribution. Specifically, the scale parameter of Gaussian or orthogonal initializations of 𝐕\mathbf{V} is chosen such that the condition number (ratio of largest to smallest singular value) of 𝐉t\mathbf{J}_{t} is bounded around 1 for all tt [2], or such that the first and second moments of the distribution of squared singular values (averaged over inputs) of the Jacobian products are chosen near 1 and 0, respectively according to dynamical isometry [4]. Parameter conditions for the latter have been derived for i.d.d. Gaussian input. The approach has been extended to RNNs under the assumption of untied weights with good empirical correspondence in vanilla and minimalRNN [3], but larger discrepancy in LSTMs [12], suggesting there are other contributions to stability in more complex, state-of-the-art architectures [6, 7].

Better understanding how robust gradient propagation emerges in these less idealized settings that include high-dimensional and temporally correlated input sequences demands a more general theory. These Jacobian products appearing in Eq. (3) can each be expressed as the transpose of a forward sequence, r=t+1s𝐉r=(𝐉s𝐉t+1)\prod_{r=t+1}^{s}\mathbf{J}^{\top}_{r}=\left(\mathbf{J}_{s}\cdots\mathbf{J}_{t+1}\right)^{\top}. Note that the backward time gradient dynamics shares properties with the forward time hidden state dynamics. This is an important insight that suggests the use of advanced tools from dynamical systems theory may help better understand gradient propagation.

Refer to caption
Figure 1: Distinguishing stability regimes of forward dynamics using Lyapunov spectra. (a-c) Example unit activations from simulations computed from 10 initial conditions and 2 input sequences (𝐱t\mathbf{x}_{t} (light) and 𝐱t\mathbf{x}_{t}^{\prime} (dark)) for variance of weight matrix 𝐕\mathbf{V} set to gives 3 examples spanning the 3 qualitatively distinct stability regimes based on the sign of λmax\lambda_{max}: stability, λmax<0\lambda_{max}<0 (a; blue, σ𝐕2=1/500\sigma^{2}_{\mathbf{V}}=1/500); marginal stability, λmax0\lambda_{max}\approx 0 (b; gray, σ𝐕2=100\sigma^{2}_{\mathbf{V}}=100); and chaos λmax>0\lambda_{max}>0 (c; green, σ𝐕2=500\sigma^{2}_{\mathbf{V}}=500). (d) Lyapunov spectra computed from (a-c) (same colors). (e) First 20 exponents from (d). (vanilla RNN with ϕ=tanh\phi=\tanh; 𝐕\mathbf{V} orthogonal; 𝐔=𝕀\mathbf{U}=\mathbb{I}; 2 realizations (light/dark) of 𝐱\mathbf{x} Gaussian, 𝚲𝐱=σ𝐱2𝕀\mathbf{\Lambda}_{\mathbf{x}}=\sigma^{2}_{\mathbf{x}}\mathbb{I} with σ𝐱2=0.6\sigma^{2}_{\mathbf{x}}=0.6; 10 Gaussian initial conditions, 𝚲𝐡0=σ𝐡02𝕀\mathbf{\Lambda}_{\mathbf{h}_{0}}=\sigma^{2}_{\mathbf{h}_{0}}\mathbb{I} with σ𝐡02=1\sigma^{2}_{\mathbf{h}_{0}}=1).

3 Background

Here we describe the stochastic Lyapunov exponents from the ergodic theory of non-autonomous dynamical systems and outline their connection to the conditions that support gradient-based learning in RNNs. Jacobians are linear maps that evolve state perturbations forward along a trajectory, 𝐮t=𝐉t𝐮t1\mathbf{u}_{t}=\mathbf{J}_{t}\mathbf{u}_{t-1}, e.g. for some initial perturbation 𝐡0(ϵ)=𝐡0+ϵ𝐮0\mathbf{h}_{0}(\epsilon)=\mathbf{h}_{0}+\epsilon\mathbf{u}_{0} with ϵ1\epsilon\ll 1 and a given vector 𝐮0\mathbf{u}_{0}. Thus, perturbations can vanish or explode under the hidden state dynamics according to the singular value spectrum of 𝐓t=𝐉t𝐉1\mathbf{T}_{t}=\mathbf{J}_{t}\cdots\mathbf{J}_{1}. The linear stability of the dynamics is obtained taking ϵ0\epsilon\to 0 and then tt\to\infty. In particular, the Lyapunov exponents, {λi}i=1N\{\lambda_{i}\}_{i=1}^{N}, are the exponential growth rates associated with the singular values of 𝐓t1/t\mathbf{T}_{t}^{1/t} for tt\to\infty, i.e. the logarithms of the eigenvalues of limt(𝐓t𝐓t)12t\lim_{t\to\infty}\left(\mathbf{T}_{t}^{\top}\mathbf{T}_{t}\right)^{\frac{1}{2t}}. As a property of the stationary dynamics, the Lyapunov exponents are independent of initial condition for ergodic systems, i.e. those with only one or the same type of attractor. If the maximum Lyapunov exponent λmax\lambda_{\textrm{max}} is positive, the stationary dynamics is chaotic and small perturbations explode, otherwise it is stable and small perturbations vanish. The theory was initially developed for autonomous dynamical systems, where stable dynamics implies limit-cycle or fixed-point attractors. How the shape of the Lyapunov spectrum varies with network model parameters has provided unprecedented insight into autonomous ( evolving in the absence of inputs) neural network models in theoretical neuroscience [13, 11, 14].

RNN dynamics are non-autonomous because the hidden state dynamics 𝐡t\mathbf{h}_{t} are driven by inputs 𝐱t\mathbf{x}_{t}. The theory of random dynamical systems generalizes stability analyses to non-autonomous dynamics driven by an input sequence sampled from a stationary distribution [15]. Analytical results typically employ uncorrelated Gaussian inputs, but the framework is expected to apply to a wider range of well-behaved input statistics. This includes those with finite, low-order moments and finite correlation times like character streams from written language and sensor data from motion capture systems. The time course of the driven dynamics depends on the specific input realization but, critically, the theory guarantees that the stationary dynamics for all input realizations share the same stability properties, which will in general depend on the input distribution (e.g., its variance). Stability here is quantified using the same definition of Lyapunov exponents in the autonomous case (now called stochastic Lyapunov exponents; we hereon will omit stochastic for brevity). In Fig. 1, we show vanilla RNN dynamics in stable, marginally stable, and chaotic regimes, as measured by the sign of the largest LE. For stable driven dynamics, the stationary activity on the time-dependent attractor (called a random sink) is independent of initial condition. This holds true also in the marginally stable case, λmax0\lambda_{max}\approx 0, where in addition there are directions along which the sizes of perturbations (error gradients) are maintained over many iterates. For chaotic driven dynamics, the activity variance over initial conditions fluctuates in time in a complex way. Understanding stability properties in this non-autonomous setting is at the frontier of analysis of neural network dynamics in both theoretical neuroscience [11, 16, 17] and machine learning [18, 19].

The practical calculation of the Lyapunov spectra (see Sec. 4) is fast and offers more robust numerical behaviour and generality over singular value decomposition. Together, this suggests the Lyapunov spectra as a useful diagnostic for RNN sequence learning.

There are a handful of features of the Lyapunov spectrum that dictate gradient propagation and thus, influence training speed and performance. The maximum LE determines the linear stability. The mean exponent determines the rate of contraction of full volume elements. The LE variance measures heterogeneity in stability across different directions and can reflect the conditioning of the product of many Jacobians. We use the first two in the analysis of our experiments later in the paper.

We remind readers that much of the theory regarding high-dimensional dynamics is derived in the so-called thermodynamic limit NN\to\infty. Here, the number of exponents diverges and the spectrum over i/Ni/N becomes stationary (so called extensive dynamics), i.e., it is insensitive to NN so long as NN is large enough (e.g. [13, 11]). Here, self-averaging effects can take hold and enable accurate analytical results based on Gaussian assumptions justified by central limit theorem arguments. ‘Large enough’ NN is typically in the hundreds (though this should be checked for any given application) and so these results can be useful for studying RNNs.

Finally, we note that obtaining the Lyapunov exponents from their definition relates to the singular values of 𝐓t1/t\mathbf{T}_{t}^{1/t} for large tt, and thus is numerically impractical. The standard approach [20] is to exploit the fact that mm-dimensional volume elements grow at a rate λ(m)=i=1mλi\lambda^{(m)}=\sum_{i=1}^{m}\lambda_{i} and so the desired rates, λ1=λ(1)\lambda_{1}=\lambda^{(1)}, λ2=λ(1)λ1\lambda_{2}=\lambda^{(1)}-\lambda_{1}, … arise from volumes obtained by projecting orthogonal to subspaces of increasing dimension. These rates are a direct output of computationally efficient orthonormalization procedures such as QR-decomposition. Next, we discuss the practical aspects of computing the Lyapunov exponents in data-driven RNNs.

4 Lyapunov Spectrum Estimation for RNNs

We extend this framework of Lyapunov exponents to recurrent neural networks by calculating the asymptotic trajectories of the hidden states of the networks when driven by the same signal.

4.1 Algorithm description

We adopt the well-established standard algorithm for computing the Lyapunov spectrum [20, 21]. We choose the driving signal to be sampled from fixed-length sequences of the test set. For each input sequence in a batch, a matrix Q is initialized as the identity, and the hidden states, h, are initialized as zero. To track the expansion and contraction of Q at each step, the Jacobian of the hidden states is calculated and then applied to a set of vectors of Q. The expansion of each vector is calculated and the matrix Q is updated using the QR decomposition at each step. If ritr_{i}^{t} is the expansion of the ithi^{th} vector at time step tt – corresponding to the ithi^{th} diagonal element of R in the QR decomposition– then the ithi^{th} Lyapunov exponent λi\lambda_{i} resulting from an input signal of length TT is given by:

λi=1Tt=1Tlog(rit)\lambda_{i}=\frac{1}{T}\sum_{t=1}^{T}{\log}(r_{i}^{t}) (5)

The Lyapunov exponents resulting from each input xjx^{j} in the batch of input sequences are calculated in parallel and then averaged. For our experiments, the Lyapunov exponents were calculated over 100 time steps with 10 different input sequences. The mean of the 10 resulting Lyapunov spectra is reported as the spectrum. An example calculation of the Lyapunov spectrum is shown in Fig. 2.

The algorithm for computing the Lyapunov exponents is described in Algorithm 1.

for 𝐱j\mathbf{x}^{j} in Batch do
       initialize h, Q
       for t = 1 \rightarrow T do
             h \leftarrow f(h, 𝐱tj\mathbf{x}_{t}^{j})
             J dfdh\leftarrow\frac{d\textbf{f}}{d\textbf{h}}
             Q \leftarrow J\cdotQ
             Q, R \leftarrow qrqr(Q)
             γi\gamma_{i} += log(Rii)
       end for
      λij=γij/T\lambda^{j}_{i}=\gamma^{j}_{i}/T
end for
λi\lambda_{i} = mean(λij)j{}_{j}(\lambda_{i}^{j})
Algorithm 1 Lyapunov Exponents Calculation
Refer to caption
(a)
Refer to caption
(b)
Figure 2: Convergence of Lyapunov spectrum estimator. (a) Lyapunov exponent estimation as a function of time, as shown in Equation (5), where TT is the number of iterations. Each line represents a different exponent in the spectrum. By 100 iterations, it is clear that each exponent is changing very little as a function of iteration number. (b) The mean Lyapunov spectrum over the set of input sequences. Standard deviation is shown by the red bars.

5 Experiments

To illustrate the computation and application of Lyapunov exponents, we applied the approach to two tasks with distinct task constraints: character prediction from sentences, where performance depends on capturing low-dimensional and long temporal correlations, and signal prediction from motion capture data, where performance relies also on signal correlations across the many dimensions of the input.

5.1 Task details

For character prediction, we use Leo Tolstoy’s War and Peace as the data and follow the character-level language modelling task outlined by Karpathy et al. in [22]. For signal prediction, we use the CMU motion-capture dataset and pre-process it using the procedure outlined in [23]. Both tasks require predicting the next step in a sequence given the preceding input sequence. For the character prediction task (CharRNN), we use input sequences of 100 characters. For the motion capture task (CMU Mocap), we use input sequences of 25 time steps.

For each task, we use a single-layer LSTM architecture with weight parameters initialized uniformly on [p,p][-p,p], where pp is referred to as the initialization parameter. For further details on the implementation of both tasks, see Supplementary Materials and in addition the Gihub repository111https://github.com/shlizee/lyapunov-hyperopt for available code and ongoing work.

5.2 Algorithm convergence properties

We find that the general shape of the spectrum is reached early in training. In Fig. 3, we use the CharRNN task as an example to show that the spectrum rapidly changes in the first few epochs of training, quickly converging after a small number of training epochs to a spectrum near that of the trained network. The reduced mean-squared difference shows a 𝒪(t1)\mathcal{O}(t^{-1}) convergence with learning epoch tt, while the mean difference shows this convergence is from below. This was true across the range of initialization parameters tested.

(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Refer to caption
(e)
Refer to caption
Figure 3: Evolution of the Lyapunov spectra over training. For values of initialization parameter given in legend in (a) (p=0.04p=0.04 (Blue), p=0.08p=0.08 (Orange), p=0.12p=0.12 (Green), p=0.16p=0.16 (Red), p=0.2p=0.2 (Purple), p=0.5p=0.5 (Brown)). The spectra before training (a) and after four (b) and fifteen (c) epochs. The reduced mean-squared error between successive epochs (d) shows that the each spectrum converges quickly during training, with t1t^{-1} plotted as a dashed black line for reference. Similarly, the difference between the mean of the Lyapunov exponents at the fifteenth epoch and the mean at previous epochs (e) rapidly approaches zero throughout training.

5.3 Performance efficiency relative to training

The calculation of the Lyapunov spectrum is very efficient relative to the computation required for training, as long as the user has an analytical expression for the Jacobian of the recurrent layer (as we describe the expressions for GRU and LSTM in supplementary materials). The average time required for calculating the spectrum for a network by evolving 10 input sequences for 100 time steps relative to the average time required to train those networks on the same device is shown in Table 1 for the two tasks we consider (c.f. section 5.1). The mean number of training units (epochs or iterations) needed to reach the minimum validation loss is calculated for each task, and the total training time is compared to the Lyapunov spectra calculation time. For both tasks, the time required to calculate the spectrum was equivalent to about 4%4\% of the total training time.

Table 1: LE calculation rate relative to training
Training Units LE Calculation Fraction of Total
Example Needed Time (training units) Training Time
CharRNN 26±1026\pm 10 epochs 1.04±.041.04\pm.04 epochs 0.04±0.020.04\pm 0.02
CMU Mocap 2.4±.5×1032.4\pm.5\times 10^{3} iters 104±3104\pm 3 iters 0.04±0.010.04\pm 0.01

Further improvements to the computation time of the exponents can be made by making slight modifications to Alg. 1. First, we can "warm-up" the hidden states h and Q onto the attractor of the dynamical system without the computational cost of the QR decomposition, allowing a faster convergence of the spectrum. A further improvement can be made by not taking the QR decomposition every time step during the calculation of the exponents, but instead every TONT_{ON} steps. Since the QR step is the most expensive computation in the algorithm, the increase in speed over orthonormalizing every step is approximately TONT_{ON}. However, since increasing this interval leads to greater expansion and contraction of the vectors of Q before orthonormalization steps, this can lead to a spurious plateau at higher indices due to the accumulation of rounding errors. However, if one cares only about the first few exponents, this effect is negligible for reasonable selections of TONT_{ON} (see supplementary materials for details about warmup and effect of increasing TONT_{ON}).

5.4 Lyapunov spectrum as a robust readout of training stability

In general, the dependence of the Lyapunov spectrum on hyperparameters is tangled. To direct our exploration of this dependence and how it relates to performance, we used the task constraints to guide our interpretation of spectra behavior of trained networks from randomly sampled hyperparameters.

For CharRNN, we hypothesized from existing work that to satisfy the constraint of propagating a scalar signal over many iterates, it would be sufficient to have a single exponent approaching 0 with the other exponents more negative [24]. Focusing here on hyperparameters, we uniformly sampled a fixed range of log-dropout and learning rate while keeping the initialization parameter fixed. In Fig. 4(a,b), we observe the spectra and indeed find a correlation between maximum Lyapunov exponent and validation loss. Networks with a maximum LE closer to zero indeed performed better. Of these 60 spectra, we selected a smaller subset of 6 that spanned the range of the gross shapes. Interestingly, they also spanned the range of losses, suggesting there is a consistent signal about the loss encoded in the complex, yet systematic variation of the maximum Lyapunov exponent with hyperparameters. The gross shape of these spectra correlated less with the loss, supporting our hypothesis. A straightforward interpretation of the signal we have observed is that the slower the activity in the mode associated with this maximum LE, the longer gradient information can propagate backward in time. This would give the dynamics more predictive power and thus better performance, but it remains to be verified.

Refer to caption
Refer to caption
Figure 4: Correlation between Lyapunov spectra and performance. Shown are spectra (a,c) and loss dependence on spectra (b,d) for networks trained on CharRNN (a,b) and CMU MoCap (c,d). Those in (a,b) are for 60 random combinations of learning rate and log-dropout rate, while those in (c,d) are for 30 values of the initialization parameter, pp. The colors in (a,b) and (c,d) are 6 samples selected by eye to roughly span the range of gross spectra shapes seen in (a) and (c), respectively. All other spectra are shown in gray.

Next, we focused on the motion capture prediction task. Here, the higher dimensional data means that the changes that training induces in the shape of the Lyapunov spectra will likely be more complicated. We hypothesized that nevertheless there would be more information in the gross shape of the spectra of the dynamics of trained networks and less in the maximum LE. Here, we vary the initialization parameter from 0.05 to 0.5 while keeping the dropout and learning rates fixed. Indeed, we find (Fig. 4c,d) consistent variation of the gross shape with loss. We have quantified the variation of loss with gross shape using the mean Lyapunov exponent, though the spectra vary with higher order features as well.

The plots of Fig. 4 illustrate that distinct properties of the LE spectrum correlate with validation performance, depending upon task structure. For the CharRNN task, the maximum exponent being larger (closer to 0), corresponds to better performance. For the Mocap learning task, the larger mean Lyapunov exponent correspond to better performance. Please see Supplementary Materials for other spectrum statistics of all networks tested.

Given our observation that the spectrum changes most rapidly early in training, we believe the most strongly-correlated metrics for a task could allow an alternative way to asses performance of a network early in training, thus serving as a hyper-parameter tuning tool.

6 Conclusion & Discussion

In this paper, we presented an exposition and example application of Lyapunov exponents for understanding training stability in RNNs. We motivated them as a natural quantity related to stability of dynamics and useful in a natural complementary approach to existing mathematical approaches for understanding training stability focused on the singular value spectrum. We adapted the standard algorithm for their computation to RNNs, and showed how it could be made more efficient in standard machine learning development environments. We demonstrated that even the basic implementation runs almost 2 orders of magnitude faster than typical training time (see [11] for further precision and efficiency considerations) and that it converges relatively quickly with learning time. The latter implies it could be useful as a readout of performance early in training. To test this application, we studied it in two tasks with distinct constraints. In both we find interpretable variation with performance reflecting these distinct constraints.

There are a few points to be made about LEs in RNNs. First, one should consider carefully the degrees of freedom in the architecture under consideration. We have considered the hidden state dynamics of an LSTM such that there are as many Lyapunov exponents as there are units. LSTMs have gating variables that can also carry their own dynamics and could be considered as degrees of freedom, adding another NN exponents to the spectrum. While we were able to obtain evidence from the spectra of hidden states alone, recent work [25, 11] suggests that studying the stability in the subspace of gating variables can also be informative. A second point is that one should also consider the role of the input weights 𝐔\mathbf{U} (c.f. Eq. 1). The strength of inputs (e.g. input signal variance) that drive high-dimensional dynamics has long been known to have a stabilizing effect [26, 16, 27]. Thus, the input statistics of the task’s data can play a role that is modulated by 𝐔\mathbf{U}, making the latter a target for gradient-based algorithms aiming to decrease loss. Due to limited space, we have not analyzed these weights here, but believe it is important to do so in applications. Finally, we raise perhaps the most glaring adaptation: that of approximating this asymptotic quantity to settings of finite-length sequences. We proposed to use averaging over inputs, in addition to averaging over initial conditions, as a way to take advantage of batch tensor mechanics to achieve faster LE estimates. This is in contrast to long-time averages used in the LE definition, although our method is justified under assumptions of ergodicity and short transient dynamics. Analyzing the influence of input batch size on this precision is a topic of future research.

The task analysis we have performed suggests a use case for the Lyapunov spectrum as having features that serve as an early readout of performance useful to speed up hyperparameter search. We looked the maximum and mean Lyapunov exponent as two features highlighted by hypotheses we made based on our knowledge of the task constraints. Presumably, there are others relevant here and in other tasks. The search for generic features that do not require knowledge of task constraints (e.g. dynamical isometry) is important, as is demonstrating how useful these features are in practise with complex sequence data. One limitation is how quickly they converge with learning. For example, as a composite quantity, the mean LE converges rather quickly, while the maximum LE is significantly slower to converge (see supplemental material). In [11], it is shown that loss function/learning rule combinations can also sculpt the spectra and alter its statistics, demonstrating that these are more sources of variation to understand. Interestingly, Full Force [28], an alternative to Back-propagation-through-time, has a qualitatively similar spectra (same max, min, and mean LE) but with much lower variance, presumably a desirable feature for generalization. This method also demands more information about performance than just a gradient, suggesting the interesting hypothesis that the precision with which the features of the Lyapunov spectra can be sculpted can scale with the amount of information provided in training. Finally, recent work has given theoretical grounding to why LSTMs avoid vanishing gradients [25]. Of course this was the reason for the design of LSTM so it only recapitulates the original design intuition. Looking forward however as more complex architectures are developed with less interpretable design, this approach based on LEs can still provide novel insight into why they work.

We close with a short discussion of open problems as this is a prominent area in understanding network dynamics at the intersection of machine learning and computational neuroscience. Having supporting analytical results is essential for robust control of complex problems. To this end, understanding how and where the assumption of untied weights breaks down and how forward propagation differs from backward propagation will be important in extending analytical results on spectral constraints. The Lyapunov spectrum can guide this work. Last, we believe insightful parallels in theoretical neuroscience work. For example, Lyapunov spectrum have been derived for additional degrees of freedom in the unit dynamics, different connectivity ensembles, and more. Also, a discrepancy between the loss of linear stability and the onset of chaotic dynamics in driven systems must be understood [27]. Making these connections explicit will serve both fields.

References

  • [1] Yoshua Bengio, Patrice Simard, and Paolo Frasconi. Learning long-term dependencies with gradient descent is difficult. IEEE transactions on neural networks, 5(2):157–166, 1994.
  • [2] Razvan Pascanu, Tomas Mikolov, and Yoshua Bengio. On the difficulty of training recurrent neural networks. In International conference on machine learning, pages 1310–1318, 2013.
  • [3] MinMin Chen, Jeffery Pennington, and Samuel. Gating Enables Signal Propagation in Recurrent Neural Networks. In ICML, 2018.
  • [4] Jeffrey Pennington, Samuel Schoenholz, and Surya Ganguli. Resurrecting the sigmoid in deep learning through dynamical isometry: theory and practice. In Advances in neural information processing systems, pages 4785–4795, 2017.
  • [5] Ben Poole, Subhaneil Lahiri, Maithra Raghu, Jascha Sohl-Dickstein, and Surya Ganguli. Exponential expressivity in deep neural networks through transient chaos. In Advances in neural information processing systems, pages 3360–3368, 2016.
  • [6] Greg Yang. Scaling limits of wide neural networks with weight sharing: Gaussian process behavior, gradient independence, and neural tangent kernel derivation. arXiv preprint arXiv:1902.04760, 2019.
  • [7] Yang Zheng and Eli Shlizerman. R-force: Robust learning for random recurrent neural networks. arXiv preprint arXiv:2003.11660, 2020.
  • [8] R Legenstein and W Maass. Edge of chaos and prediction of computational performance for neural circuit models. Neural Networks, 20(3):323–334, 2007.
  • [9] Jeffrey Pennington, Samuel S Schoenholz, and Surya Ganguli. The emergence of spectral universality in deep networks. arXiv preprint arXiv:1802.09979, 2018.
  • [10] Thomas Laurent and James von Brecht. A recurrent neural network without chaos. arXiv preprint arXiv:1612.06212, 2016.
  • [11] Rainer Engelken, Fred Wolf, and LF Abbott. Lyapunov spectra of chaotic recurrent neural networks. arXiv preprint arXiv:2006.02427, 2020.
  • [12] Dar Gilboa, Bo Chang, Minmin Chen, Greg Yang, Samuel S Schoenholz, Ed H Chi, and Jeffrey Pennington. Dynamical isometry and a mean field theory of lstms and grus. arXiv preprint arXiv:1901.08987, 2019.
  • [13] Michael Monteforte and Fred Wolf. Dynamical entropy production in spiking neuron networks in the balanced state. Phys. Rev. Lett., 105(26):1–4, dec 2010.
  • [14] Maximilian Puelma Touzel. Cellular dynamics and stable chaos in balanced networks. PhD thesis, Georg-August University, 2015.
  • [15] Ludwig Arnold. Random Dynamical Systems. Springer, 1991.
  • [16] Guillaume Lajoie, Kevin K. Lin, and Eric Shea-Brown. Chaos and reliability in balanced spiking networks with temporal drive. Phys. Rev. E, 87(5):1–5, 2013.
  • [17] Guillaume Hennequin, Tim Vogels, and Wulfram Gerstner. Non-normal amplification in random balanced neuronal networks. Phys. Rev. E, 86(1):1–12, jul 2012.
  • [18] Giancarlo Kerg, Kyle Goyette, Maximilian Puelma Touzel, Gauthier Gidel, Eugene Vorontsov, Yoshua Bengio, and Guillaume Lajoie. Non-normal recurrent neural network (nnrnn): learning long time dependencies while improving expressivity with transient dynamics. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems 32, pages 13613–13623. Curran Associates, Inc., 2019.
  • [19] Guan-Horng Liu and Evangelos A Theodorou. Deep learning theory review: An optimal control and dynamical systems perspective. arXiv preprint arXiv:1908.10920, 2019.
  • [20] Giancarlo Benettin, Luigi Galgani, Antonio Giorgilli, and Jean-Marie Strelcyn. Lyapunov characteristic exponents for smooth dynamical systems and for hamiltonian systems; a method for computing all of them. part 1: Theory. Meccanica, 15(1):9–20, 1980.
  • [21] L Dieci and E S Van Vleck. Computation of a few Lyapunov exponents for continuous and discrete dynamical systems. Applied Numerical Mathematics, 17(3):275–291, 1995.
  • [22] Andrej Karpathy, Justin Johnson, and Li Fei-Fei. Visualizing and understanding recurrent networks. arXiv preprint arXiv:1506.02078, 2015.
  • [23] Chen Li, Zhen Zhang, Wee Sun Lee, and Gim Hee Lee. Convolutional sequence to sequence model for human dynamics. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2018.
  • [24] Mikael Henaff, Arthur Szlam, and Yann LeCun. Recurrent orthogonal networks and long-memory tasks. arXiv preprint arXiv:1602.06662, 2016.
  • [25] Tankut Can, Kamesh Krishnamurthy, and David J Schwab. Gating creates slow modes and controls phase-space complexity in grus and lstms. arXiv preprint arXiv:2002.00025, 2020.
  • [26] L. Molgedey, J. Schuchhardt, and H. G. Schuster. Suppressing chaos in neural networks by noise. Phys. Rev. Lett., 69(26):3717–3719, 1992.
  • [27] Jannis Schuecker, Sven Goedeke, and Moritz Helias. Optimal Sequence Memory in Driven Random Networks. Physical Review X, 8(4):41029, 2018.
  • [28] Brian DePasquale, Christopher J Cueva, Kanaka Rajan, G Sean Escola, and LF Abbott. full-force: A target-based method for training recurrent networks. PloS one, 13(2):e0191527, 2018.

Appendix A Task Implementation Details

A.1 CharRNN: Character Prediction Task

For our CharRNN experiment, the text used for training the networks was Leo Tolstoy’s War and Peace, which contains about 3.2 million characters of mostly English text, with a total of 82 distinct characters in the vocabulary. We split the data into train, validation, and test sets with 80/10/10 ratio, respectively. We study the performance of single-layer LSTM networks (with no bias vectors) with hidden size of 512 units trained on this data set. We train the network using stochastic gradient descent with a batch size of 100 and the Adam optimizer. The network is unrolled for 100 time steps. After 10 epochs of training, the learning rate is decreased by a multiplying by a factor a .95 after each additional epoch. The batch size used for training was 100.

For each combination of dropout and learning rate, the network was trained for 40 epochs, and the training epoch at which the network had the lowest validation loss was selected.

A.2 CMU Mocap: Signal Prediction Task

The CMU Motion Capture (CMU Mocap) dataset is a high-dimensional data set capturing the location of various sensors on a subject over time as they conduct a variety of tasks. In particular, we narrow our focus onto the Sports and Activities category of the data set. The basketball, soccer, walking, and jumping subcategories were selected to compile a variety of motions which require consistent full-body movement (unlike, for example, the "directing traffic" and "car washing" sub-categories). For this task, we use a standard 256-unit LSTM.

We chose the training and validation split for the data corresponding to each task as 80/20. After this split, the data for all tasks considered was mixed into a single training and validation set. For each dimension of the data, the mean and the standard deviation across all times points is calculated, and those dimensions for which the standard deviation is less than 10410^{-4} are ignored. All other dimensions are normalized by their standard deviation. For our example, we consider the first 20 features of the remaining data. The network is trained to predict the 20-dimensional output at the time step tt given the preceding input sequence. The loss function is defined as mean-squared error across the 20 dimensional-output. We chose the batch size of 250 for training.

A.3 Trained Network Hyperparameters

The following are the network parameters for the networks highlighted for each task in Figure 4.

Table 2: CharRNN Networks Hyperparameters
Learning Initialization
Color Dropout Rate Parameter, p
Blue 0.034 0.00301 0.08
Orange 0.000 0.00319 0.08
Green 0.039 0.00574 0.08
Red 0.028 0.00055 0.08
Purple 0.014 0.00153 0.08
Brown 0.106 0.00485 0.08
Table 3: CMU Mocap Networks Hyperparameters
Learning Initialization
Color Dropout Rate Parameter, p
Blue 0.0 0.05 0.476
Orange 0.0 0.05 0.305
Green 0.0 0.05 0.196
Red 0.0 0.05 0.309
Purple 0.0 0.05 0.059
Brown 0.0 0.05 0.439

Appendix B Analytical Jacobian Expressions

We aim to calculate the derivative of the hidden states hh of the recurrent neural network at time tt with respect to the hidden states at the previous time step, t1t-1,

htht1.\frac{\partial h_{t}}{\partial h_{t-1}}.

When the network has multiple recurrent layers, we can partition the hidden state hh the hidden states at layer kk. Then, if a given recurrent network has ll layers,

ht=[ht1ht2htkhtl]h_{t}=\left[\begin{matrix}h_{t}^{1}\\ h_{t}^{2}\\ \vdots\\ h_{t}^{k}\\ \vdots\\ h_{t}^{l}\end{matrix}\right]

We will denote the input of layer kk as xtkx^{k}_{t}. We can easily use the equations for a given recurrent network type to calculate the derivative of the hidden state of a layer kk, htkh_{t}^{k} with respect to the hidden states of that same layer at the previous time step htkh_{t}^{k}, as well as with respect to the external input into that layer xtkx^{k}_{t}. The expressions for htkht1k\frac{\partial h^{k}_{t}}{\partial h^{k}_{t-1}} and htkxtk\frac{\partial h^{k}_{t}}{\partial x^{k}_{t}} for LSTMs and GRUs are given in the following sections.
Since layer kk takes the output of layer k1k-1 as input, xtk=htk1x^{k}_{t}=h_{t}^{k-1} for k2k\geq 2. Then, in order to find the derivative of the kthk^{th} layer with respect to the a previous layer,

htkht1j,jk,\frac{\partial h^{k}_{t}}{\partial h^{j}_{t-1}},\quad j\leq k,

we need to compose the derivative with the derivative with respect to the input for each preceding layer until layer jj is reached, at which point we calculate the derivative with respect to the hidden state at the previous time step, htjht1j\frac{\partial h^{j}_{t}}{\partial h^{j}_{t-1}}. This gives the composition:

htkht1j=htkhtk1htk1htk2htj+1htjhtjht1j\frac{\partial h^{k}_{t}}{\partial h^{j}_{t-1}}=\frac{\partial h^{k}_{t}}{\partial h^{k-1}_{t}}\cdot\frac{\partial h^{k-1}_{t}}{\partial h^{k-2}_{t}}\cdot...\cdot\frac{\partial h^{j+1}_{t}}{\partial h^{j}_{t}}\cdot\frac{\partial h^{j}_{t}}{\partial h^{j}_{t-1}}

For each htihti1,j<ik\frac{\partial h^{i}_{t}}{\partial h^{i-1}_{t}},j<i\leq k, we use the expression for htixti\frac{\partial h^{i}_{t}}{\partial x^{i}_{t}}.

When implementing this in matrix form, we start by calculating the derivatives for the first layer, then each layer below it. The composition described above can be taken by multiplying by the block directly above it. This reduces redundant computation and requires only a loop as long as the number of layers in the RNN. Given the structure of one-directional recurrent layers, this matrix would be block lower-triangular, with the size of each block corresponding to the number of hidden units in the layer.

B.1 LSTM Derivatives

The LSTM is defined by the equations

ft=σg(Wfxt+Ufht1+bf)\displaystyle f_{t}=\sigma_{g}\left(W_{f}x_{t}+U_{f}h_{t-1}+b_{f}\right)
it=σg(Wixt+Uiht1+bi)\displaystyle i_{t}=\sigma_{g}\left(W_{i}x_{t}+U_{i}h_{t-1}+b_{i}\right)
ot=σg(Woxt+Uoht1+bo)\displaystyle o_{t}=\sigma_{g}\left(W_{o}x_{t}+U_{o}h_{t-1}+b_{o}\right)
ct=σh(Wcxt+Ucht1+bc),\displaystyle c_{t}=\sigma_{h}\left(W_{c}x_{t}+U_{c}h_{t-1}+b_{c}\right),

where σg\sigma_{g} is the sigmoid function and σh\sigma_{h} is tanhtanh (both applied element-wise). We use the following notation when calculating these derivatives:

y=Wxt+Uht1+b,y_{*}=W_{*}x_{t}+U_{*}h_{t-1}+b_{*},

where is determined by the gate/state. For LSTM, there are 4 different gates/states, (f,i,o,cf,~{}i,~{}o,~{}c), each with a corresponding y,W,U,by_{*},W_{*},U_{*},b_{*}. We also use \circ to indicate elementwise multiplication (if both vectors) or row-wise multiplication (if vector and matrix).
In other words, if v,un,Wn×n\textbf{v},\textbf{u}\in\mathcal{R}^{n},\textbf{W}\in\mathcal{R}^{n\times n}, then

uv=[u1v1ukvkunvn],uW=[u1W11u1W1ju1W1nukWk1ukWkjukWknunWn1unWnjunWnn]\displaystyle\textbf{u}\circ\textbf{v}=\left[\begin{matrix}u_{1}\cdot v_{1}\\ \vdots\\ u_{k}\cdot v_{k}\\ \vdots\\ u_{n}\cdot v_{n}\end{matrix}\right],\quad\textbf{u}\circ\textbf{W}=\left[\begin{matrix}u_{1}\cdot W_{11}&\ldots&u_{1}\cdot W_{1j}&\ldots&u_{1}\cdot W_{1n}\\ \vdots&&\vdots&&\vdots\\ u_{k}\cdot W_{k1}&\ldots&u_{k}\cdot W_{kj}&\ldots&u_{k}\cdot W_{kn}\\ \vdots&&\vdots&&\vdots\\ u_{n}\cdot W_{n1}&\ldots&u_{n}\cdot W_{nj}&\ldots&u_{n}\cdot W_{nn}\\ \end{matrix}\right]

The derivatives of each of these gates/states with respect to the hidden states and the input is shown below. The final line shows what the derivative of the hidden state is relative to the hidden state at the previous time step.

ftht1=[σ(yf)(1σ(yf))]TUf\displaystyle\frac{\partial f_{t}}{\partial h_{t-1}}=\left[\sigma(y_{f})\circ\left(1-\sigma(y_{f})\right)\right]^{T}\circ U_{f}
itht1=[σ(yi)(1σ(yi))]TUi\displaystyle\frac{\partial i_{t}}{\partial h_{t-1}}=\left[\sigma(y_{i})\circ\left(1-\sigma(y_{i})\right)\right]^{T}\circ U_{i}
otht1=[σ(yo)(1σ(yo))]TUo\displaystyle\frac{\partial o_{t}}{\partial h_{t-1}}=\left[\sigma(y_{o})\circ\left(1-\sigma(y_{o})\right)\right]^{T}\circ U_{o}
ctht1=ftht1ct1+itht1tanh(yc)+itsech2(yc)Uc\displaystyle\frac{\partial c_{t}}{\partial h_{t-1}}=\frac{\partial f_{t}}{\partial h_{t-1}}\circ c_{t-1}+\frac{\partial i_{t}}{\partial h_{t-1}}\circ\text{tanh}(y_{c})+i_{t}\circ\text{sech}^{2}(y_{c})\circ U_{c}
htht1=otht1tanh(ct)+otsech2(ct)ctht1\displaystyle\frac{\partial h_{t}}{\partial h_{t-1}}=\frac{\partial o_{t}}{\partial h_{t-1}}\circ\text{tanh}(c_{t})+o_{t}\circ\text{sech}^{2}(c_{t})\circ\frac{\partial c_{t}}{\partial h_{t-1}}
ftxt=[σ(yf)(1σ(yf))]TWf\displaystyle\frac{\partial f_{t}}{\partial x_{t}}=\left[\sigma(y_{f})\circ\left(1-\sigma(y_{f})\right)\right]^{T}\circ W_{f}
itxt=[σ(yi)(1σ(yi))]TWi\displaystyle\frac{\partial i_{t}}{\partial x_{t}}=\left[\sigma(y_{i})\circ\left(1-\sigma(y_{i})\right)\right]^{T}\circ W_{i}
otxt=[σ(yo)(1σ(yo))]TWo\displaystyle\frac{\partial o_{t}}{\partial x_{t}}=\left[\sigma(y_{o})\circ\left(1-\sigma(y_{o})\right)\right]^{T}\circ W_{o}
ctxt=ftxtct1+itxttanh(yc)+itsech2(yc)Wc\displaystyle\frac{\partial c_{t}}{\partial x_{t}}=\frac{\partial f_{t}}{\partial x_{t}}\circ c_{t-1}+\frac{\partial i_{t}}{\partial x_{t}}\circ\text{tanh}(y_{c})+i_{t}\circ\text{sech}^{2}(y_{c})\circ W_{c}
htxt=otxttanh(ct)+otsech2(ct)ctxt\displaystyle\frac{\partial h_{t}}{\partial x_{t}}=\frac{\partial o_{t}}{\partial x_{t}}\circ\text{tanh}(c_{t})+o_{t}\circ\text{sech}^{2}(c_{t})\circ\frac{\partial c_{t}}{\partial x_{t}}

Appendix C LE calculation alternatives and their effects

The algorithm for calculation of LE can be altered by introducing two alterations. (c.f. Algo. 2). First, we can ‘warm-up’ the hidden states h and the vectors of Q by evolving them via the network dynamics for a set number of steps before calculating the Lyapunov Exponents. This allows the hidden states and vectors of Q to relax onto the attractor of the dynamical system defined by the network. This means that all samples used are from the stationary distribution on the attractor. Without the warmup, the initial samples are from a transient phase with a distribution over initial conditions that lies off the attractor and so has a bias effect on the average for finite sample averaging.

(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
Figure 5: The effect of warmup on the calculation Lyapunov spectrum. Shown are the Lyanpunov exponents as a function of iteration with (a) no warmup and warmups of (b) 100 time steps and (c) 250 time steps. It is clear that the case with no warmup has large changes early in the evolution of the spectrum, whereas those with warmup periods have much smaller changes early in the evolution.

A further improvement can be made by not taking the QR decomposition every time step during the calculation of the exponents, but instead every TONT_{ON} steps. Since the QR step is the most expensive computation in the algorithm, the increase in speed over orthonormalizing every step is approximately TONT_{ON}. However, since increasing this interval leads to greater expansion and contraction of the vectors of Q before orthonormalization steps, this can lead to a spurious plateau at higher indices due to the accumulation of rounding errors (c.f. [11]. However, if one cares only about the first few exponents, this effect is negligible for reasonable selections of TONT_{ON}. In Fig. 6, we show how increasing TONT_{ON} flattens out the spectrum beyond a certain index, but that the beginning of the spectrum remains unaffected by the selection of TONT_{ON}.

(a)
Refer to caption
(b)
Refer to caption
Figure 6: The effect of selection of TONT_{ON} on the resulting spectrum. Shown are (a) the full Lyapunov spectrum and (b) the first half of the spectrum for a trained network with different choices of TONT_{ON} (part of full spectrum plot indicated by gray box). Whereas the selection of TONT_{ON} clearly impacts the second half of the spectrum, the first half remains nearly identical for all these selections of TONT_{ON}.
for xjx^{j} in Ensemble do
       initialize h, Q
       Warm-up h, Q
       for t = 1 \rightarrow T do
             h \leftarrow f(h, xtjx^{j}_{t})
             J dfdh\leftarrow\frac{d\textbf{f}}{d\textbf{h}}
             Q \leftarrow J\cdotQ
             if t % TON=0T_{ON}=0 then
                   Q, R \leftarrow qrqr(Q)
                   γi\gamma_{i} += log(Rii)
             end if
            
       end for
      λi=γi/T\lambda_{i}=\gamma_{i}/T
end for
λi\lambda_{i} = mean(λij)j{}_{j}(\lambda_{i}^{j})
Algorithm 2 Modified Lyapunov Exponents Calculation

Appendix D Spectrum metrics’ correlation with loss

(a) CharRNN
Refer to caption
(b) CMU Mocap
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 7: Lyapunov spectrum statistics against validation loss for both tasks. The results for the CharRNN task are shown on the left and CMU Mocap are shown on the right. Each row represents a different statistic of the spectrum.