This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Visualizing convolutional neural network for classifying gravitational waves from core-collapse supernovae

Seiya Sasaoka Department of Physics, Tokyo Institute of Technology, 2-12-1 Ookayama, Meguro-ku, Tokyo 152-8551, Japan    Naoki Koyama Graduate School of Science and Technology, Niigata University, 8050 Ikarashi-2-no-cho, Nishi-ku, Niigata City, Niigata 950-2181, Japan    Diego Dominguez Department of Physics, Tokyo Institute of Technology, 2-12-1 Ookayama, Meguro-ku, Tokyo 152-8551, Japan    Yusuke Sakai Department of Design and Data Science and Research Center for Space Science, Advanced Research Laboratories,
Tokyo City University, 3-3-1 Ushikubo-Nishi, Tsuzuki-ku, Yokohama, Kanagawa 224-8551, Japan
   Kentaro Somiya Department of Physics, Tokyo Institute of Technology, 2-12-1 Ookayama, Meguro-ku, Tokyo 152-8551, Japan    Yuto Omae Artificial Intelligence Research Center, College of Industrial Technology, Nihon University, 1-2-1 Izumi-cho, Narashino, Chiba 275-8575, Japan    Hirotaka Takahashi Department of Design and Data Science and Research Center for Space Science, Advanced Research Laboratories,
Tokyo City University, 3-3-1 Ushikubo-Nishi, Tsuzuki-ku, Yokohama, Kanagawa 224-8551, Japan
Institute for Cosmic Ray Research (ICRR), The University of Tokyo, 5-1-5 Kashiwa-no-Ha, Kashiwa City, Chiba 277-8582, Japan Earthquake Research Institute, The University of Tokyo, 1-1-1 Yayoi, Bunkyo-ku, Tokyo 113-0032, Japan
Abstract

In this study, we employ a CNN to classify GWs originating from CCSNe. Training is conducted using spectrograms derived from three-dimensional numerical simulations of waveforms, which are injected onto real noise data from the O3 observation of both Advanced LIGO and Advanced Virgo. To gain insights into the decision-making process of the model, we apply CAM techniques to visualize the regions in the input image that are significant for the model’s prediction. The CAM maps reveal that the model’s predictions predominantly rely on specific features within the input spectrograms, namely the gg-mode and low-frequency modes. The visualization of convolutional neural network models provides interpretability to enhance their reliability and offers guidance for improving detection efficiency.

I Introduction

The first detection of GWs from a binary black hole merger by Advanced LIGO [1] in 2015 marked the beginning of GW astronomy [2]. Throughout three observing runs (O1, O2, and O3), Advanced LIGO and Advanced Virgo [3] reported 90 GW events [4, 5, 6, 7]. As of May 2023, the international GW network, now including KAGRA [8], has begun its fourth observing run (O4) with improved sensitivity.

All of the GW events detected so far are exclusively from compact binary coalescences. However, short-duration GW bursts arising from CCSNe are expected to be detected by the current and next-generation GW detectors, such as the Einstein Telescope [9] and the Cosmic Explorer [10]. CCSNe, resulting from massive star explosions leading to neutron stars or stellar-mass black holes, stand as one of the most energetic astrophysical events in the Universe, emitting electromagnetic waves, neutrinos, and GWs. While electromagnetic waves from CCSNe are frequently observed, neutrinos have only been detected from SN1987A [11, 12]. GWs are expected to carry information about the inner core’s dynamics, providing vital insights into the explosion mechanism which remains elusive. The primary conundrum lies in discerning how a stalled shock wave is revived to cause a star to explode. Currently, there are two prevailing theories [13]: the neutrino-driven mechanism [14], in which shock waves are revived by neutrinos stored behind the shock wave heating the surrounding matter, and the magnetorotational mechanism [15], in which the rapid rotation of the progenitor causes explosions driven by strong magnetic fields. The typical GW detection range for neutrino-driven signals is expected to be around 10 kpc, while the detection range for magnetorotational signals is expected to be above 100 kpc [16].

Due to the stochastic nature of GW signals from CCSNe, the conventional matched-filtering technique, which relies on specific waveform templates, is unsuitable. Alternative detection methods based on the time-frequency representation have been devised in response. In particular, the coherent WaveBurst (cWB) pipeline [17, 18] detects and reconstructs burst GW signals by searching for excess power in a time-frequency map, with minimal reliance on a specific source model.

Predicting GW signals from CCSNe remains a formidable challenge. However, recent advancements in theoretical research and multidimensional numerical simulations have revealed certain signal properties. For neutrino-driven CCSNe, the dominant emissions arise from the gg-mode oscillation of the proto-neutron star (PNS) surface. These frequencies progressively increase over time, ranging from a few hundred Hz to a few kHz. Additionally, at low frequencies (\lesssim 200 Hz), GW emissions associated with hydrodynamics instabilities including neutrino-driven convection and standing accretion shock instability (SASI) [19] are observed in some simulations. Insights obtained from these simulations are pivotal in enhancing methods for CCSNe detection and analysis.

In recent years, machine learning techniques, especially deep learning, have gained traction in a variety of scientific fields due to their capacity for recognizing intricate patterns and extracting meaningful features from large data sets. This ability has been especially noted in areas such as computer vision and natural language processing. Its application in GW research has followed, with numerous implementations and explorations, as highlighted in a comprehensive review in Ref. [20] and the foundational efforts by George and Huerta [21, 22]. In the field of CCSNe analysis, Astone et al. [23] leveraged CNNs to detect CCSNe within Gaussian noise, using gg-mode phenomenological waveforms for training, outperforming the cWB pipeline. Consecutive studies by Iess et al. [24, 25] involved the training of both one- and two-dimensional CNNs and long short-term memory networks [26] to identify seven distinct CCSN waveforms embedded in real noise and glitches, with their models achieving 98% classification accuracy at 1 kpc with a three-detector network. Additionally, Chan et al. [27] employed one-dimensional CNNs to investigate both magnetorotational and neutrino-driven signals in Gaussian noise, recording a true alarm probability of 80% for magnetorotational signals from sources at 60 kpc and 55% for neutrino-driven signals from sources at 10 kpc with a fixed false alarm probability of 10%. In another study, Edwards [28] used two-dimensional CNNs to classify 18 different equations of state (EOS) from pure magnetorotational CCSN signals, attaining an accuracy of 72%. López et al. [29] refined phenomenological waveforms originally used by Astone et al. [23], achieving a 60% true alarm probability for signals located at 15 kpc with a 5% false alarm rate.

Although deep learning exhibits strong performance on a wide range of tasks, its intricate models, characterized by a large number of parameters, pose challenges in elucidating their decision-making processes. To address this, the field of explainable artificial intelligence [30] has surged, aiming to make model decisions transparent and interpretable. Within the context of CNNs, efforts have been made to develop techniques that attempt to understand the decision-making process by reverse mapping the output of the network into the input space to identify the specific input components that were discriminative in producing the output. Class activation mapping (CAM) [31] is one such method, which computes a weighted sum of the outputs of the last convolutional layer using the outputs of the global average pooling layer after the last convolutional layer as weights. It helps identify the regions in the input image that were important for a prediction, but the model needs to be modified to include a global average pooling layer, which may result in lower accuracy. Gradient-weighted class activation mapping (Grad-CAM) [32] was introduced as a solution to this limitation of CAM, offering the advantage of not requiring any modifications to the network architecture by using gradient information from the prediction for weighted parameters. Subsequently, Grad-CAM++  [33], a generalization of Grad-CAM, and Score-CAM [34], a gradient-free CAM method, were developed to generate more accurate saliency maps than Grad-CAM. These techniques to analyze deep learning models are commonly used in fields such as electrocardiogram signal analysis [35] and x-ray diagnosis [36]; however, for GW analysis, they have only been used in Ref. [37] to the best of our knowledge.

In this study, we first take an approach similar to Ref. [25] and train a two-dimensional CNN model to classify CCSNe signals using short-time Fourier-transformed spectrograms as input for simplicity. We use nine types of waveforms from recent three-dimensional numerical simulations and O3 real noise to train and validate our model. In the test, signals from sources between 1 and 10 kpc are considered, and the performance of the model for sources at each distance is discussed. To interpret the model, we use three CAM methods to generate saliency maps and evaluate them using two metrics: average drop and average increase. The best CAM method is then applied to correctly classified and also misclassified samples to visualize the regions in the input spectrogram that influence the predictions of our model.

The remainder of this chapter is organized as follows. Section II describes our data sets, the CNN model, and the CAM techniques. In Sec. III, we discuss the classification performance of our model, and apply multiple visualization techniques to interpret the model. We summarize and conclude the paper in Sec. IV.

II Method

Our CNN model is trained to classify strains at the three detectors LIGO Hanford (H1), LIGO Livingston (L1), and Virgo (V1) into ten classes: noise and nine different CCSN waveforms. In this section, we first provide an overview of the data used in this study, including a brief summary of the CCSN simulation data and the preprocessing strategy to generate our training, validation, and test sets. Subsequently, our CNN architecture and the theory of the visualization technique of the model are explained.

II.1 Data set

II.1.1 CCSN waveforms

Modeling the stellar core collapse, bounce, and subsequent post-bounce evolution is very complicated and computationally expensive. However, remarkable advancements in three-dimensional numerical simulations of neutrino-driven explosions have been achieved by several groups in recent years. Specific details of the waveform depend on various properties of the progenitor, such as the mass, angular velocity, and EOS of the dense matter. Both the general relativity approximation and the handling of neutrino transport critically influence simulations. From the available simulation data under a variety of conditions, we select nine types of waveforms from four recent three-dimensional numerical simulations [38, 39, 40, 41]. All of them allow us to compute the GW amplitude in any observer direction from the quadrupole moment.

Powell and Müller [38] performed simulations using the general-relativistic neutrino hydrodynamics code CoCoNuT-FMT [42]. We use two waveforms from the models he3.5 and s18. The progenitor of he3.5 is an ultra-stripped star evolved from a helium star with an initial mass of 3.5 MM_{\odot}. The simulation is stopped at 0.7 s after core bounce. The GW is dominated by excitation of gg-modes in the PNS with a peak frequency around 900 Hz. Model s18 is a single star with a zero-age main-sequence (ZAMS) mass of 18 MM_{\odot}. The simulation was stopped 0.89 s after core bounce. The GW emission is similar to model he3.5, with gg-modes oscillations of the PNS with a peak frequency around 900 Hz.

Radice et al. [39] studied eight models using the Eulerian radiation-hydrodynamics code FORNAX [43]. We use waveforms from the models s13 and s25 corresponding to progenitors of 13 and 25 MM_{\odot} ZAMS, respectively. The simulation is ended at 0.77 s in s13 and 0.62 s in s25 after bounce. Both waveforms are characterized by ff- and gg- modes with a peak frequency around 1400 Hz in s13 and 1100 Hz in s25. In addition, the s25 waveform has a clear SASI mode around 100 Hz.

From the simulation by Powell and Müller [40], we use the three models m39, y20, and s18np. Model m39 is a rapidly rotating 39 MM_{\odot} Wolf-Rayet star with an initial surface rotation velocity of 600 kms1\mathrm{km}\ \mathrm{s}^{-1}. It produces a neutrino-driven explosion without magnetic fields. The other two are nonrotating models of 20 MM_{\odot} Wolf-Rayet star and an 18 MM_{\odot} ZAMS star. The simulation is ended at 0.98, 1.2, 0.56 s after core bounce in models m39, y20, and s18np, respectively. All three models show GW emission associated with prompt convection shortly after bounce as well as ff-mode oscillations of the PNS. In model s18np, the absence of strong perturbations from convective oxygen burning, in contrast to s18, prevents the shock from being revived and leads to the development of strong SASI activity, with a frequency reaching \sim 400 Hz by the end of the simulation.

Powell et al. [41] performed simulations using three different EOSs: LS220 [44], SFHx, and SFHo [45]. The progenitor models are 85 and 100 MM_{\odot} Population III ZAMS stars. We use the z85_sfhx and z100_sfho models. The simulation is ended at 0.59 s in z85_sfhx and 0.62 s in z100_sfho. Both waveforms show typical gg-mode emission with a peak frequency of \sim 700 Hz. In the z85_sfhx model, the frequency of the SASI emission increases up to the point of shock revival, reaching \sim 200 Hz, and decreases afterwards. In the z100_sfho model, which does not explode, the frequency of the SASI emission continues to increase by the end of the simulation, reaching \sim 400 Hz.

Figure 1 shows the amplitude spectral density of the plus mode of each waveform at 1 kpc from the polar direction. The amplitude of the m39 waveform is the largest of these waveforms. Waveforms that have peaks around 100 Hz such as s25 and s13 indicate that they have SASI-induced GW modes.

Refer to caption
Figure 1: Amplitude spectral density of the plus mode of each waveform at 1 kpc. The observer is in the polar direction.

II.1.2 Data processing

From the simulation data presented in the previous section, we calculate the amplitude of the GW for generating the data sets. The directions of radiation (θ,ϕ)(\theta,\phi) are uniformly sampled and the plus and cross polarizations of each GW are calculated using the formulas

h+\displaystyle h_{+} =1D2Gc4(Q¨θθQ¨ϕϕ),\displaystyle=\frac{1}{D}\frac{2G}{c^{4}}(\ddot{Q}_{\theta\theta}-\ddot{Q}_{\phi\phi}), (1)
h×\displaystyle h_{\times} =1DGc4Q¨θϕ,\displaystyle=\frac{1}{D}\frac{G}{c^{4}}\ddot{Q}_{\theta\phi}, (2)

where QQ is the traceless quadrupole moment and DD is the distance between a source and Earth. As the sampling of the simulation is usually not uniform in time, we resample data uniformly with a sampling rate of 4096 Hz. A high-pass filter with a cutoff frequency of 11 Hz and a Tukey window with α=0.1\alpha=0.1 are applied to the resampled signals. Each signal is then truncated or padded with zeros to make the length 1 second. In order to make the model robust, we randomly time shift the signals so that the time of the core bounce is between 0 and 0.15 s. For the training and validation sets, the signals are scaled using an optimal matched filter signal-to-noise ratio (SNR), as the amplitudes of the simulated signals are quite different. The SNR is defined as

ρ=4fminfmax|h~(f)|2Sn(f)df,\rho=\sqrt{4\int_{f_{\min}}^{f_{\max}}\frac{|\tilde{h}(f)|^{2}}{S_{n}(f)}\mathrm{d}f}, (3)

where h~(f)\tilde{h}(f) is the Fourier transform of the signal and Sn(f)S_{n}(f) is the one-side power spectral density of the noise. The network SNR of the detectors H1, L1, and V1, given by

ρnet=ρH12+ρL12+ρV12,\rho_{\mathrm{net}}=\sqrt{\rho_{\mathrm{H1}}^{2}+\rho_{\mathrm{L1}}^{2}+\rho_{\mathrm{V1}}^{2}}, (4)

is used to scale the signals. We generate samples with network SNRs from 20 to 50 for training and validation sets. For the test set, the signals are scaled to have distances between 1 and 10 kpc. Sky location is also randomly selected and the GW amplitude h(t)h(t) is computed, taking into account the antenna pattern functions F+F_{+} and F×F_{\times} and the delay in arrival time of each detector with the following equation:

h(t)\displaystyle h(t) =F+(α,δ,ψ,t)h+(t+Δt)\displaystyle=F_{+}(\alpha,\delta,\psi,t)h_{+}(t+\Delta t)
+F×(α,δ,ψ,t)h×(t+Δt),\displaystyle\qquad+F_{\times}(\alpha,\delta,\psi,t)h_{\times}(t+\Delta t), (5)

where α\alpha is the right ascension, δ\delta is the declination, and ψ\psi is the polarization angle. Δt\Delta t is the delay in arrival time between the detector and the center of the Earth. We use the PyCBC software library [46] to carry out these computations.

Noise used in this study is O3 real data of Advanced LIGO and Advanced Virgo, obtained from the Gravitational Wave Open Science Center [47]. Data from GPS time 1238236470 to 1238252308 is used for the training set, 1238265720 to 1238354855 is used for the validation set, and 1238404064 to 1238457121 is used for the test set. Data around the event time reported in the second Gravitational Wave Transient Catalog [6] are excluded. After a signal is injected into the noise, each sample is whitened with the power spectral density computed using Welch’s method [48] and then short-time Fourier transformed with a window size of 0.0625 seconds to produce a spectrogram. The spectrogram is normalized to [0, 1] before input to the network.

We generated 60 000 samples for the training and validation sets, and 100 000 samples for the test set. The test set has 1000 samples for each class and each distance. Sample spectrograms in the training set are shown in Fig. 2.

Refer to caption
Figure 2: Sample whitened spectrograms of each class at the H1 detector in the training set. Each signal sample is observed in the polar direction and scaled to have an SNR of 40. The bounce time is fixed at 0.1 s.

II.2 CNN model

Our CNN model consists of two convolutional layers of kernel size 3, each followed by a max-pooling layer of size 2 and a rectified linear unit (ReLU) layer. The outputs of these layers are fed into two fully connected layers, and finally the softmax layer outputs a size-10 vector whose elements represent a probability of each class. The model has 427 378 trainable parameters in total. This model is shallower than the one used in Ref. [25]. However, its classification performance is comparable to the previous study, prompting us to adopt this model. Reducing the number of layers also helps us generate higher-resolution CAM maps.

The model is trained using categorical cross entropy as the loss function and Adam optimizer [49] with a learning rate of 5×1045\times 10^{-4} to update the weights. In the training, we adopt curriculum learning [50] as a strategy to enhance the model and accelerate the training by starting from inputting high-SNR samples and gradually adding lower-SNR samples. We train the model on a single GPU (NVIDIA GeForce RTX3090) for 120 epochs with a mini-batch size of 128.

II.3 Visualization

After training the model, we use CAM techniques to generate saliency maps. These maps show the regions in the input that influenced the model’s prediction. In this study, we select three CAM methods—Grad-CAM, Grad-CAM++, and Score-CAM—which are widely used today to interpret CNN models. All of these CAM techniques are applied to the convolutional layer prior to the final max-pooling layer in our model.

II.3.1 Grad-CAM

Grad-CAM is a gradient-based visualization technique that highlights the important regions of an input image that the model is looking at while making a prediction. Suppose that for a given input, the prediction score for class cc before the softmax layer of the trained model is ycy^{c}, and the kkth output matrix of the last convolutional layer is AkA^{k}. To obtain the Grad-CAM map of class cc, we first compute the gradients of the score ycy^{c} with respect to the (i,j)(i,j) component of the kkth feature map AkA^{k}. We then take the global average of these gradients:

αkc=1Zi,jycAijk,\alpha_{k}^{c}=\frac{1}{Z}\sum_{i,j}\partialderivative{y^{c}}{A_{ij}^{k}}, (6)

where ZZ is the number of pixels in AkA^{k}. This weight αkc\alpha_{k}^{c} represents the importance of the feature map kk for the class cc.

The Grad-CAM map of the class cc is computed as a linear sum of AkA^{k} with αkc\alpha_{k}^{c} as weights. The ReLU function is applied to extract only features that have a positive contribution to the prediction score. The resulting map of class cc is expressed as

LGradCAMc=ReLU(kαkcAk).L_{\mathrm{Grad-CAM}}^{c}=\mathrm{ReLU}\pqty{\sum_{k}\alpha_{k}^{c}A^{k}}. (7)

Since convolutional layers and pooling layers make the size of the feature map smaller than the input, the Grad-CAM map is finally interpolated to make it the same size as the input.

II.3.2 Grad-CAM++

While Grad-CAM takes a global average of the gradient matrix when calculating the weight αkc\alpha_{k}^{c} in Eq. (6), Chattopadhay et al. [33] proposed a method to fully include the importance of each pixel in the gradient matrix by taking its weighted average for the weight:

αkc=i,jαijkcReLU(ycAijk).\alpha_{k}^{c}=\sum_{i,j}\alpha_{ij}^{kc}\ \mathrm{ReLU}\pqty{\partialderivative{y^{c}}{A_{ij}^{k}}}. (8)

The ReLU function is used to account for features that increase the activation of the output neuron rather than suppress the activation of the output neuron. The weights αijkc\alpha_{ij}^{kc} can be theoretically derived using higher-order derivatives:

αijkc=2yc(Aijk)222yc(Aijk)2+a,bAabk3yc(Aijk)3.\alpha_{ij}^{kc}=\frac{\frac{\partial^{2}y^{c}}{(\partial A_{ij}^{k})^{2}}}{2\frac{\partial^{2}y^{c}}{(\partial A_{ij}^{k})^{2}}+\sum_{a,b}A_{ab}^{k}\frac{\partial^{3}y^{c}}{(\partial A_{ij}^{k})^{3}}}. (9)

This method is known as Grad-CAM++, since it can be considered as a generalization of Grad-CAM. The saliency map for Grad-CAM++ is expressed in the same way as for Grad-CAM, using weights in Eq. (8) and feature maps, as

LGradCAM++c=ReLU(kαkcAk).L_{\mathrm{Grad-CAM++}}^{c}=\mathrm{ReLU}\pqty{\sum_{k}\alpha_{k}^{c}A^{k}}. (10)

II.3.3 Score-CAM

Wang et al. [34] proposed a gradient-free CAM method called Score-CAM. It solves the problem of gradient-based CAM methods, namely that the gradient is unstable, easily disturbed by noise, and can vanish or explode in deep networks. To generate a Score-CAM map, feature maps are used to mask an input image. Let HkH^{k} be the kkth feature map, up-sampled to the same size as the input and normalized to [0, 1]. Given an input image XX, the weight for the kkth feature map is computed as the difference between the score of the masked image XHkX\circ H^{k} and the score of the baseline image XbX_{\mathrm{b}}:

αk=f(XHk)f(Xb),\alpha_{k}=f(X\circ H^{k})-f(X_{\mathrm{b}}), (11)

where f()f(\cdot) denotes the output of the CNN and \circ denotes the Hadamard product. A black image is used as a baseline image. The Score-CAM map of class cc is then computed as a linear sum of the ccth value of αk\alpha_{k} and the feature map AkA^{k} as

LScoreCAMc=ReLU(kαkcAk).L_{\mathrm{Score-CAM}}^{c}=\mathrm{ReLU}\pqty{\sum_{k}\alpha_{k}^{c}A^{k}}. (12)

III Results and Discussion

III.1 Classification performance

Figure 3 shows the evolution of the categorical cross entropy loss function during the training. Initially, we input samples with SNRs between 40 and 50. In the subsequent 40 epochs, samples with SNRs between 30 and 40 are input, leading to the temporary increase in the loss at epoch 40. In the last 40 epochs, as we input samples with SNRs between 20 and 30, a similar temporary increase in loss is observed at epoch 80. From the loss curves, we confirm that there is no significant overfitting to the training data.

Refer to caption
Figure 3: Loss curves for the training and validation sets. During the first 40 epochs, we input samples with SNRs between [40, 50], followed by [30, 40] SNR samples in the subsequent 40 epochs, and [20, 30] SNR samples in the last 40 epochs.
Refer to caption
Figure 4: True positive rate of each waveform in the test set against source distance.
Refer to caption
(a) 1 kpc
Refer to caption
(b) 5 kpc
Refer to caption
(c) 10 kpc
Figure 5: Confusion matrices of the test set (left) and violin plots of network SNR of each waveform (right) from sources with distances of 1, 5, and 10 kpc.

Classification accuracy is defined as the proportion of correctly classified samples out of the total number of samples. After training, the model achieves a classification accuracy of 97.8% on a validation set consisting of uniformly sampled signals with SNRs between 20 and 50. On the test set, our model shows an accuracy of 98.4% for signals with sources from 1 kpc, which is comparable to the results of the previous study [25], despite some differences in the condition that we used O3 noise instead of O2 noise and performed ten-class classification instead of eight-class classification. In Fig. 4, we plot a true positive rate (TPR) for each waveform in the test set against distance. A TPR, also known as the sensitivity of a class cc, is defined as the ratio of the number of samples correctly classified into class cc to the number of samples of class cc in the test set. For signals from sources at 1 kpc, each waveform has a TPR greater than 90%, and this decreases monotonically with the distance of the source, having an average TPR of 26.1% at 10 kpc. For the m39 waveform, because the amplitude of the strain is much larger than the others due to its rapid rotation and high explosion energy, the TPR for sources at 10 kpc is 99.2%.

The performance of a multiclass classifier is also expressed by a confusion matrix, which shows the number of samples classified into each class. Figure 5 plots the confusion matrices normalized for each class and the distribution of the network SNR for signals from sources at 1, 5, and 10 kpc. We can see from the confusion matrices that as the distance increases, the amplitude of the signal becomes smaller and the number of samples misclassified as noise increases. The accuracy for signals at 10 kpc is 33.2%, and our model cannot identify most of these signals, except for the m39 waveforms, whose SNR is much higher than others with a median value of 47.9.

III.2 Dimensionality reduction

Before implementing CAM techniques, we used the t-distributed stochastic neighbor embedding (t-SNE) [51] algorithm to see if the convolutional layers in the model can extract the features in the input to classify samples. The t-SNE algorithm is a dimensionality-reduction technique that minimizes the Kullback-Leibler divergence between two probability distributions: one representing pairwise similarities between data points in the original high-dimensional space and another representing pairwise similarities in a lower-dimensional space. In our CNN model, each sample is compressed into a vector with a length of 2112 before the dense layers. The t-SNE algorithm is used to map this vector into two-dimensional space to make it interpretable for humans. We visualize the dimensionally reduced feature maps of the test set, whose signals are coming from sources at 1 kpc, for which our model shows a good classification accuracy. The visualized data are shown in Fig. 6. We can clearly see that there are ten clusters in the data set and our model could extract meaningful features to classify these samples into ten classes. The fact that some signal samples are also found in the noise cluster and that s13 samples are found in other clusters, especially in the noise cluster, is consistent with the results of the confusion matrix in Fig. 5(a).

Refer to caption
Figure 6: Features of the test samples at 1 kpc extracted by CNN and mapped into two-dimensional space by the t-SNE algorithm.

III.3 Saliency maps

To quantitatively evaluate different CAM methods, we use two metrics—average drop and average increase [33]—which focus on the change in a model’s score caused by the explanation map. An explanation map for a target class cc is generated as element-wise multiplication of a saliency map LcL^{c} with an original image XX:

Ec=LcX.E^{c}=L^{c}\circ X. (13)

Average drop measures the percentage decrease in a model’s score for a target class cc when inputting only the explanation map, instead of the original image. It is expressed as

Averagedrop=1001Nimax(0,yicoic)yic,\mathrm{Average\ drop}=100\cdot\frac{1}{N}\sum_{i}\frac{\max(0,y_{i}^{c}-o_{i}^{c})}{y_{i}^{c}}, (14)

where yicy_{i}^{c} is the score for class cc on the iith original image and oico_{i}^{c} is the score on the explanation map. The lower this value, the more effective the visualization method, since the explanation map includes more of the relevant information for making a correct prediction.

Average increase measures the number of samples in the data set, and the model’s confidence increases when providing only the explanation map as input. It is expressed as

Averageincrease=1001NiΘ(oicyic),\mathrm{Average\ increase}=100\cdot\frac{1}{N}\sum_{i}\Theta(o_{i}^{c}-y_{i}^{c}), (15)

where Θ\Theta is the Heaviside step function. Unlike the previous metric, the higher this value is, the more effective the visualization method will be because there are more samples that score higher when given the explanation map than when given the original image.

For the three visualization methods Grad-CAM, Grad-CAM++, and Score-CAM, the two metrics described above are computed using signals from sources at 1 kpc in the test set. The results are summarized in Table 1. Score-CAM shows the best results in both metrics, meaning that it is the best visualization technique for our model among the three CAM methods considered in this study. We also qualitatively compare these methods by visualizing some samples. One example is shown in Fig. 7. The input image is represented by a color image, with the red, green, and blue channels corresponding to the H1, L1, and V1 spectrograms, respectively. All three saliency maps take large values around the SASI mode around 100 Hz. At high frequencies, the Grad-CAM and Grad-CAM++ maps only take slightly larger values around 1 kHz, whereas Score-CAM has gg-mode-like arch shapes around 1 kHz. This suggests that the visualization by Score-CAM captures more of the input features that are discriminative for the prediction.

Table 1: Results for evaluation of the explanations generated by Grad-CAM, Grad-CAM++, and Score-CAM on the test set.
Method Grad-CAM Grad-CAM++ Score-CAM
Ave. Drop (%)
(Lower is better)
30.70 17.58 9.61
Ave. Increase (%)
(Higher is better)
1.30 1.40 1.96
Refer to caption
Figure 7: Qualitative comparison of three CAM maps for the s25 sample at 1 kpc.
Refer to caption
Figure 8: Input spectrograms and Score-CAM maps of correctly classified samples. The SNR of each signal sample is 40.

As discussed above, we determine that the Score-CAM is the optimal method for generating saliency maps for our model. We produce saliency maps by Score-CAM for the inputs of each class, which can be seen in Fig. 8. In the input images, as in the previous figure, the red, green, and blue channels corresponding to the H1, L1, and V1 data, respectively. This means, for example, that in the reddish image such as the m39 sample in Fig. 8, the SNR at the H1 detector is smaller than that at the L1 and V1 detectors. All of the plotted signal samples are scaled to have an SNR of 40 and are correctly classified by our model. We plotted several CAM maps for noise samples in addition to the one in this figure, but the regions that the model sees to predict them to be noise were random. In the he3.5 and s13 samples, we can see that the model focuses on the gg-mode arch shape, especially in their low- and high- frequencies areas. In the s18 and y20 models, the models see all of the gg-modes. In the s18np model, the CAM map indicates that the model considers not only the gg-mode but also prompt convection and SASI. The s25 model has SASI activity, but its amplitude is not too large, and the CAM map shows that the model’s prediction is based on the prompt convection and the high-frequency gg-mode. In the m39, z85_sfhx, and z100_sfho models, the CAM maps take large values at high frequencies in the gg-mode. In addition, in the z85_sfhx and z100_sfho models, the low-frequency SASI mode, whose frequency increases with time, is also visible in the CAM maps. To summarize these outcomes, we found that the model looks at the gg-mode in all signal waveforms, and also looks at SASI and prompt convection in some signal waveforms when classified.

Additionally, we plot saliency maps of the misclassified samples. Figure 9 shows a spectrogram of the s25 signal sample and the Score-CAM map, which the model classified as s18np. An example spectrogram of the correct class s18np is also shown. The SNR of this signal is 85, which is quite large, and the gg-mode and the prompt convection are visible, but there is a glitch in the strain at the L1 detector. The Score-CAM map shows that the model focuses on the prompt convection and the glitch, which are used to determine that the signal is s18np. Because of this glitch, the model predicted the signal as s18np, whose gg-mode frequency increases in a shorter period of time. Another example is plotted in Fig. 10. This sample contains a s13 signal with an SNR of 48, and there are no glitches, but the model classified it as y20. We can see a SASI-induced GW mode around 100 Hz from 0.2 to 0.4 s, but the Score-CAM map indicates that the model only looks at a portion of the gg-mode and does not see the low-frequency mode.

From the misclassified samples and the Score-CAM maps, it is found that the performance of the model is sometimes affected by glitches, and does not fully take advantage of the characteristics of the signals. The former could be resolved by generating training sets that contain more glitches, and the latter could be resolved by using a time-frequency representation that is better able to reflect the various features of the CCSN signals.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Figure 9: s25 sample classified as s18np. (a) Input spectrogram. The red, green, and blue channels correspond to the H1, L1, and V1 data, respectively. There is a glitch in the L1 data. (b) Score-CAM map. (c) Example spectrogram of a s18np sample.
Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Figure 10: s13 sample classified as y20. (a) Input spectrogram. The red, green, and blue channels correspond to the H1, L1, and V1 data, respectively. (b) Score-CAM map. (c) Example spectrogram of a y20 sample.

IV Conclusions

In this study, we trained a two-dimensional CNN model to classify CCSN GW signals immersed in real noise of O3 observation data. Our model showed a comparable result to the previous study [25] for signals from sources with distances of 1 kpc. To interpret the trained model, we used the t-SNE algorithm and mapped the extracted features by the convolutional layers into a two-dimensional space. The dimension-reduced features show that the convolutional filters could extract meaningful features that are significant for classifying the signals. To gain insights into the decision-making process of the model, we applied the CAM technique to visualize the regions in the inputs that were influential to the predictions. Three methods—Grad-CAM, Grad-CAM++, and Score-CAM—were considered and we concluded that Score-CAM is the best for our model in terms of the average drop and average increase metrics. The Score-CAM maps of correctly classified signal samples revealed that the model’s predictions were heavily affected by a part of the entire gg-mode in the spectrogram of each signal. In some waveform models such as s18np, s25, z85_sfho, and z100_sfho, their CAM maps suggest that the prompt convection or SASI-induced GW mode also affects the model’s prediction.

It is important to note that \sim 4% of the pure noise test samples are identified as signal, which means that our model produces a false alarm every \sim 25 s, making it unsuitable as a detection pipeline. Since this study is the first to focus on the interpretability of CNN models in GW data analysis and serves as a first step in showcasing the effectiveness of the CAM techniques, we did not prioritize its viability as a detection methodology. To utilize machine learning models for a future detection pipeline, it is crucial to lower the false alarm rate. Using the CAM techniques can potentially enhance the efficacy of CNN models for this purpose.

In this analysis, a time-frequency map was created from the short-time Fourier transform, but its resolution is limited by the uncertainty relationship between time and frequency. In future studies, we would like to improve the accuracy of the CNN model by using methods such as the Hilbert-Huang transform [52], which can generate higher-resolution time-frequency maps, and to confirm that the CNN can also utilize several more GW modes to classify CCSN signals.

Acknowledgements.
The authors would like to thank Jade Powell for providing gravitational-wave simulation data. This research was supported in part by JSPS Grant-in-Aid for Scientific Research [No. 22H01228 (K. Somiya), and Nos. 19H01901, 23H01176 and 23H04520 (H. Takahashi)]. This research was also supported by the Joint Research Program of the Institute for Cosmic Ray Research, University of Tokyo and Tokyo City University Prioritized Studies. This research has made use of data or software obtained from the Gravitational Wave Open Science Center (gwosc.org), a service of the LIGO Scientific Collaboration, the Virgo Collaboration, and KAGRA. This material is based upon work supported by NSF’s LIGO Laboratory which is a major facility fully funded by the National Science Foundation, as well as the Science and Technology Facilities Council (STFC) of the United Kingdom, the Max-Planck-Society (MPS), and the State of Niedersachsen/Germany for support of the construction of Advanced LIGO and construction and operation of the GEO600 detector. Additional support for Advanced LIGO was provided by the Australian Research Council. Virgo is funded, through the European Gravitational Observatory (EGO), by the French Centre National de Recherche Scientifique (CNRS), the Italian Istituto Nazionale di Fisica Nucleare (INFN) and the Dutch Nikhef, with contributions by institutions from Belgium, Germany, Greece, Hungary, Ireland, Japan, Monaco, Poland, Portugal, Spain. KAGRA is supported by Ministry of Education, Culture, Sports, Science and Technology (MEXT), Japan Society for the Promotion of Science (JSPS) in Japan; National Research Foundation (NRF) and Ministry of Science and ICT (MSIT) in Korea; Academia Sinica (AS) and National Science and Technology Council (NSTC) in Taiwan.

References

  • Aasi et al. [2015] J. Aasi et al. (LIGO Scientific Collaboration), Advanced LIGO, Classical and Quantum Gravity 32, 074001 (2015).
  • Abbott et al. [2016] B. P. Abbott et al. (LIGO Scientific Collaboration and Virgo Collaboration), Observation of gravitational waves from a binary black hole merger, Phys. Rev. Lett. 116, 061102 (2016).
  • Acernese et al. [2014] F. Acernese et al., Advanced Virgo: a second-generation interferometric gravitational wave detector, Classical and Quantum Gravity 32, 024001 (2014).
  • Abbott et al. [2019] B. P. Abbott et al. (LIGO Scientific Collaboration and Virgo Collaboration), GWTC-1: A gravitational-wave transient catalog of compact binary mergers observed by LIGO and Virgo during the first and second observing runs, Phys. Rev. X 9, 031040 (2019).
  • Abbott et al. [2021a] R. Abbott et al. (LIGO Scientific Collaboration and Virgo Collaboration), GWTC-2: Compact binary coalescences observed by LIGO and Virgo during the first half of the third observing run, Phys. Rev. X 11, 021053 (2021a).
  • Abbott et al. [2021b] R. Abbott et al. (LIGO Scientific Collaboration and Virgo Collaboration), GWTC-2.1: Deep extended catalog of compact binary coalescences observed by LIGO and Virgo during the first half of the third observing run (2021b), arXiv:2108.01045 [gr-qc] .
  • Abbott et al. [2023a] R. Abbott et al. (LIGO Scientific Collaboration, Virgo Collaboration, and KAGRA Collaboration), GWTC-3: Compact binary coalescences observed by LIGO and Virgo during the second part of the third observing run, Phys. Rev. X 13, 041039 (2023a).
  • Akutsu et al. [2019] T. Akutsu et al., KAGRA: 2.5 generation interferometric gravitational wave detector, Nature Astronomy 3, 35 (2019).
  • Punturo et al. [2010] M. Punturo et al., The Einstein Telescope: a third-generation gravitational wave observatory, Classical and Quantum Gravity 27, 194002 (2010).
  • Abbott et al. [2017] B. P. Abbott et al., Exploring the sensitivity of next generation gravitational wave detectors, Classical and Quantum Gravity 34, 044001 (2017).
  • Bionta et al. [1987] R. M. Bionta, G. Blewitt, C. B. Bratton, D. Casper, A. Ciocio, R. Claus, B. Cortez, M. Crouch, S. T. Dye, S. Errede, G. W. Foster, W. Gajewski, K. S. Ganezer, M. Goldhaber, T. J. Haines, T. W. Jones, D. Kielczewska, W. R. Kropp, J. G. Learned, J. M. LoSecco, J. Matthews, R. Miller, M. S. Mudan, H. S. Park, L. R. Price, F. Reines, J. Schultz, S. Seidel, E. Shumard, D. Sinclair, H. W. Sobel, J. L. Stone, L. R. Sulak, R. Svoboda, G. Thornton, J. C. van der Velde, and C. Wuest, Observation of a neutrino burst in coincidence with supernova 1987A in the large magellanic cloud, Phys. Rev. Lett. 58, 1494 (1987).
  • Hirata et al. [1987] K. Hirata, T. Kajita, M. Koshiba, M. Nakahata, Y. Oyama, N. Sato, A. Suzuki, M. Takita, Y. Totsuka, T. Kifune, T. Suda, K. Takahashi, T. Tanimori, K. Miyano, M. Yamada, E. W. Beier, L. R. Feldscher, S. B. Kim, A. K. Mann, F. M. Newcomer, R. Van, W. Zhang, and B. G. Cortez, Observation of a neutrino burst from the supernova SN1987A, Phys. Rev. Lett. 58, 1490 (1987).
  • Janka [2012] H.-T. Janka, Explosion mechanisms of core-collapse supernovae, Annual Review of Nuclear and Particle Science 62, 407 (2012).
  • Bethe and Wilson [1985] H. A. Bethe and J. R. Wilson, Revival of a stalled supernova shock by neutrino heating, Astrophys. J.  295, 14 (1985).
  • LeBlanc and Wilson [1970] J. M. LeBlanc and J. R. Wilson, A Numerical Example of the Collapse of a Rotating Magnetized Star, Astrophys. J.  161, 541 (1970).
  • Szczepańczyk et al. [2021] M. J. Szczepańczyk, J. M. Antelis, M. Benjamin, M. Cavaglià, D. Gondek-Rosińska, T. Hansen, S. Klimenko, M. D. Morales, C. Moreno, S. Mukherjee, G. Nurbek, J. Powell, N. Singh, S. Sitmukhambetov, P. Szewczyk, O. Valdez, G. Vedovato, J. Westhouse, M. Zanolin, and Y. Zheng, Detecting and reconstructing gravitational waves from the next galactic core-collapse supernova in the advanced detector era, Phys. Rev. D 104, 102002 (2021).
  • Klimenko and Mitselmakher [2004] S. Klimenko and G. Mitselmakher, A wavelet method for detection of gravitational wave bursts, Classical and Quantum Gravity 21, S1819 (2004).
  • Klimenko et al. [2016] S. Klimenko, G. Vedovato, M. Drago, F. Salemi, V. Tiwari, G. A. Prodi, C. Lazzaro, K. Ackley, S. Tiwari, C. F. Da Silva, and G. Mitselmakher, Method for detection and reconstruction of gravitational wave transients with networks of advanced detectors, Phys. Rev. D 93, 042004 (2016).
  • Blondin et al. [2003] J. M. Blondin, A. Mezzacappa, and C. DeMarino, Stability of standing accretion shocks, with an eye toward core-collapse supernovae, The Astrophysical Journal 584, 971 (2003).
  • Cuoco et al. [2020] E. Cuoco, J. Powell, M. Cavaglià, K. Ackley, M. Bejger, C. Chatterjee, M. Coughlin, S. Coughlin, P. Easter, R. Essick, H. Gabbard, T. Gebhard, S. Ghosh, L. Haegel, A. Iess, D. Keitel, Z. Márka, S. Márka, F. Morawski, T. Nguyen, R. Ormiston, M. Pürrer, M. Razzano, K. Staats, G. Vajente, and D. Williams, Enhancing gravitational-wave science with machine learning, Machine Learning: Science and Technology 2, 011002 (2020).
  • George and Huerta [2018a] D. George and E. A. Huerta, Deep neural networks to enable real-time multimessenger astrophysics, Phys. Rev. D 97, 044039 (2018a).
  • George and Huerta [2018b] D. George and E. Huerta, Deep learning for real-time gravitational wave detection and parameter estimation: Results with Advanced LIGO data, Physics Letters B 778, 64 (2018b).
  • Astone et al. [2018] P. Astone, P. Cerdá-Durán, I. Di Palma, M. Drago, F. Muciaccia, C. Palomba, and F. Ricci, New method to observe gravitational waves emitted by core collapse supernovae, Phys. Rev. D 98, 122002 (2018).
  • Iess et al. [2020] A. Iess, E. Cuoco, F. Morawski, and J. Powell, Core-collapse supernova gravitational-wave search and deep learning classification, Machine Learning: Science and Technology 1, 025014 (2020).
  • Iess et al. [2023] A. Iess, E. Cuoco, F. Morawski, C. Nicolaou, and O. Lahav, LSTM and CNN application for core-collapse supernova search in gravitational wave real data, A&A 669, A42 (2023).
  • Hochreiter and Schmidhuber [1997] S. Hochreiter and J. Schmidhuber, Long short-term memory, Neural Computation 9, 1735 (1997).
  • Chan et al. [2020] M. L. Chan, I. S. Heng, and C. Messenger, Detection and classification of supernova gravitational wave signals: A deep learning approach, Phys. Rev. D 102, 043022 (2020).
  • Edwards [2021] M. C. Edwards, Classifying the equation of state from rotating core collapse gravitational waves with deep learning, Phys. Rev. D 103, 024025 (2021).
  • López et al. [2021] M. López, I. Di Palma, M. Drago, P. Cerdá-Durán, and F. Ricci, Deep learning for core-collapse supernova detection, Phys. Rev. D 103, 063011 (2021).
  • Arrieta et al. [2019] A. B. Arrieta, N. Díaz-Rodríguez, J. D. Ser, A. Bennetot, S. Tabik, A. Barbado, S. García, S. Gil-López, D. Molina, R. Benjamins, R. Chatila, and F. Herrera, Explainable artificial intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI (2019), arXiv:1910.10045 [cs.AI] .
  • Zhou et al. [2016] B. Zhou, A. Khosla, A. Lapedriza, A. Oliva, and A. Torralba, Learning deep features for discriminative localization, in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (IEEE Computer Society, Los Alamitos, CA, USA, 2016) pp. 2921–2929.
  • Selvaraju et al. [2017] R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and D. Batra, Grad-CAM: Visual explanations from deep networks via gradient-based localization, in 2017 IEEE International Conference on Computer Vision (ICCV) (2017) pp. 618–626.
  • Chattopadhay et al. [2018] A. Chattopadhay, A. Sarkar, P. Howlader, and V. N. Balasubramanian, Grad-CAM++++: Generalized gradient-based visual explanations for deep convolutional networks, in 2018 IEEE Winter Conference on Applications of Computer Vision (WACV) (IEEE, 2018).
  • Wang et al. [2020] H. Wang, Z. Wang, M. Du, F. Yang, Z. Zhang, S. Ding, P. Mardziel, and X. Hu, Score-CAM: Score-weighted visual explanations for convolutional neural networks, in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (IEEE Computer Society, Los Alamitos, CA, USA, 2020) pp. 111–119.
  • Jahmunah et al. [2022] V. Jahmunah, E. Ng, R.-S. Tan, S. L. Oh, and U. R. Acharya, Explainable detection of myocardial infarction using deep learning models with grad-cam technique on ecg signals, Computers in Biology and Medicine 146, 105550 (2022).
  • Panwar et al. [2020] H. Panwar, P. Gupta, M. K. Siddiqui, R. Morales-Menendez, P. Bhardwaj, and V. Singh, A deep learning and grad-CAM based color visualization approach for fast detection of COVID-19 cases using chest X-ray and CT-Scan images, Chaos, Solitons & Fractals 140, 110190 (2020).
  • Abbott et al. [2022] T. C. Abbott, E. Buffaz, N. Vieira, M. Cabero, D. Haggard, A. Mahabal, and J. McIver, GWSkyNet-Multi: A machine-learning multiclass classifier for LIGO–Virgo public alerts, The Astrophysical Journal 927, 232 (2022).
  • Powell and Müller [2019] J. Powell and B. Müller, Gravitational wave emission from 3D explosion models of core-collapse supernovae with low and normal explosion energies, Monthly Notices of the Royal Astronomical Society 487, 1178 (2019).
  • Radice et al. [2019] D. Radice, V. Morozova, A. Burrows, D. Vartanyan, and H. Nagakura, Characterizing the gravitational wave signal from core-collapse supernovae, The Astrophysical Journal Letters 876, L9 (2019).
  • Powell and Müller [2020] J. Powell and B. Müller, Three-dimensional core-collapse supernova simulations of massive and rotating progenitors, Monthly Notices of the Royal Astronomical Society 494, 4665 (2020).
  • Powell et al. [2021] J. Powell, B. Müller, and A. Heger, The final core collapse of pulsational pair instability supernovae, Monthly Notices of the Royal Astronomical Society 503, 2108 (2021).
  • Müller et al. [2010] B. Müller, H.-T. Janka, and H. Dimmelmeier, A NEW MULTI-DIMENSIONAL GENERAL RELATIVISTIC NEUTRINO HYDRODYNAMIC CODE FOR CORE-COLLAPSE SUPERNOVAE. i. METHOD AND CODE TESTS IN SPHERICAL SYMMETRY, The Astrophysical Journal Supplement Series 189, 104 (2010).
  • Skinner et al. [2019] M. A. Skinner, J. C. Dolence, A. Burrows, D. Radice, and D. Vartanyan, Fornax: A flexible code for multiphysics astrophysical simulations, The Astrophysical Journal Supplement Series 241, 7 (2019).
  • Lattimer and Swesty [1991] J. M. Lattimer and F. D. Swesty, A generalized equation of state for hot, dense matter, Nuclear Physics A 535, 331 (1991).
  • Steiner et al. [2013] A. W. Steiner, M. Hempel, and T. Fischer, CORE-COLLAPSE SUPERNOVA EQUATIONS OF STATE BASED ON NEUTRON STAR OBSERVATIONS, The Astrophysical Journal 774, 17 (2013).
  • Nitz et al. [2023] A. Nitz, I. Harry, D. Brown, C. M. Biwer, J. Willis, T. D. Canton, C. Capano, T. Dent, Larne Pekowsky, G. S. C. Davies, Soumi De, M. Cabero, Shichao Wu, A. R. Williamson, D. Macleod, B. Machenschalk, F. Pannarale, Prayush Kumar, S. Reyes, Dfinstad, S. Kumar, M. Tápai, L. Singer, Veronica-Villa, S. Khan, P. Kumar, S. Fairhurst, Bhooshan Uday Varsha Gadre, Maxtrevor, and Koustav Chandra, gwastro/pycbc: v2.2.2 release of pycbc (2023).
  • Abbott et al. [2023b] R. Abbott et al. (The LIGO Scientific Collaboration, the Virgo Collaboration, and the KAGRA Collaboration), Open data from the third observing run of LIGO, Virgo, KAGRA, and GEO, The Astrophysical Journal Supplement Series 267, 29 (2023b).
  • Welch [1967] P. Welch, The use of fast fourier transform for the estimation of power spectra: A method based on time averaging over short, modified periodograms, IEEE Transactions on Audio and Electroacoustics 15, 70 (1967).
  • Kingma and Ba [2017] D. P. Kingma and J. Ba, Adam: A method for stochastic optimization (2017), arXiv:1412.6980 [cs.LG] .
  • Bengio et al. [2009] Y. Bengio, J. Louradour, R. Collobert, and J. Weston, Curriculum learning, in Proceedings of the 26th Annual International Conference on Machine Learning, ICML ’09 (Association for Computing Machinery, New York, NY, USA, 2009) p. 41–48.
  • van der Maaten and Hinton [2008] L. van der Maaten and G. Hinton, Visualizing data using t-SNE, Journal of Machine Learning Research 9, 2579 (2008).
  • Takeda et al. [2021] M. Takeda, Y. Hiranuma, N. Kanda, K. Kotake, T. Kuroda, R. Negishi, K. Oohara, K. Sakai, Y. Sakai, T. Sawada, H. Takahashi, S. Tsuchida, Y. Watanabe, and T. Yokozawa, Application of the Hilbert-Huang transform for analyzing standing-accretion-shock-instability induced gravitational waves in a core-collapse supernova, Phys. Rev. D 104, 084063 (2021).