This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Deep Learning and Explainable AI: New Pathways to Genetic Insights

Chenyu WangContributed equally to this work.    Chaoying Zuofootnotemark:    Zihan Sufootnotemark:    Yuhang Xing    Lu Li    Maojun Wang***Corresponding authors: Maojun Wang (Email:mjwang@mail.hzau.edu.cn) and Zeyu Zhang (Email:zhangzeyu@mail.hzau.edu.cn).    Zeyu Zhang**footnotemark: * National Key Laboratory of Crop Genetic Improvement, Hubei Hongshan Laboratory, Huazhong Agricultural University, 430070, Hubei, China Agricultural College, Shihezi University, 832002, Xinjiang, China
Abstract

Deep learning-based AI models have been extensively applied in genomics, achieving remarkable success across diverse applications. As these models gain prominence, there exists an urgent need for interpretability methods to establish trustworthiness in model-driven decisions. For genetic researchers, interpretable insights derived from these models hold significant value in providing novel perspectives for understanding biological processes. Current interpretability analyses in genomics predominantly rely on intuition and experience rather than rigorous theoretical foundations. In this review, we systematically categorize interpretability methods into input-based and model-based approaches, while critically evaluating their limitations through concrete biological application scenarios. Furthermore, we establish theoretical underpinnings to elucidate the origins of these constraints through formal mathematical demonstrations, aiming to assist genetic researchers in better understanding and designing models in the future. Finally, we provide feasible suggestions for future research on interpretability in the field of genetics.

\paperid

123

1 Introduction

Deep learning has demonstrated powerful modeling and analytical capabilities in 3D genomics and regulatory genomics, enabling efficient mining of complex regulatory patterns from high-dimensional genomic data [17]. In 3D genomics, deep learning models can predict spatial features of chromatin, including dynamic changes of topologically associating domains (TADs) [42], chromatin compartments [42], and chromatin loop [19], as shown in Figure 1(a), decoding genome function from sequence through structure [11]. These models can also decipher cell-type-specific chromosomal spatial organization patterns [30], quantify spatial interaction strengths between genomic loci, and uncover their associations with gene expression regulation. In regulatory genomics, deep learning facilitates the integration of multi-omics data (e.g., Hi-C [18], ATAC-seq [16], HiChIP [26]) to predict regulatory variants of DNA methylation [43], predict potential motif-motif interactions [2, 10], efficiently identify cis-regulatory elements [39] and infer cooperative transcription factor binding networks [10], as presented in Figure 1(b, c).

Refer to caption
Figure 1: Applications of deep learning in 3D genomes and re- gulatory genomes. a. Deep learning models can predict spatial structural features of chromatin, such as chromatin compartments and topologically associating domains (TADs). b. Deep learning is also effective in predicting chromosome accessibility and histone modifications (e.g., H3K27ac). c. Additional applications include identifying cis-regulatory elements (e.g., enhancers/silencers), pre- dicting transcription factor binding motifs, and inferring precise transcription factor binding sites.

Although deep learning methods have achieved remarkable success in these biological fields, due to the use of multi-layer nonlinear transformations, complex architectures, and a large number of parameters, deep learning models are often perceived as “ black boxes” by humans, making their internal mechanisms difficult to interpret and understand. The inherent “ black-box ” nature of these models not only limits their reliability and practical applicability, but also makes it hard to understand how their predictions relate to actual biological processes, which hinders the translation of research findings into biological mechanisms [3].

In recent years, the development of eXplainable AI (XAI) methods has emerged as a promising approach to address the “ black-box” challenge [25]. These methods aim to reveal the inner workings of neural network models in ways that are understandable to humans, or to explain how specific input features influence the prediction of model, thereby enhancing model transparency and trustworthiness. Therefore, many recent studies have employed interpretable deep learning methods to gain biologically relevant insights, such as identifying cis-regulatory elements that influence gene expression [34]. In this work, we systematically review and categorize the current interpretability techniques accordingly applied to two major tasks: 3D genome modeling and gene regulatory network prediction.

Despite various interpretable deep learning methods have shown promising results, each comes with its own limitations, and there remains a lack of rigorous mathematical proofs regarding the specific limitations of their applicability. In particular, many critical claims in existing studies (e.g.,“ enforcing transparency techniques may compromise model performance” [28]) are primarily based on empirical observations and lack solid theoretical foundations. To address this gap, we provide theoretical analyses of the limitations of certain interpretability methods.

The contributions of this paper are as follows:

• We propose a novel classification framework that categorizes interpretable deep learning methods in genomics into two types: (1) Input interpretability, encompassing convolutional kernel visualization, gradient-based methods, and perturbation-based methods. (2) Model interpretability, including attention mechanisms and transparent models based on biological prior knowledge.

• We not only provide intuitive explanations of each method and introduce their applications but also rigorously reveal the inherent limitations of selected interpretable approaches through mathematical derivation, thus offering theoretical support for method evaluation and selection.

2 Interpretable methods

In this section, we systematically introduce interpretable approaches in deep learning applications for genomics, categorizing them into two main types: input interpretability and model interpretability. Input interpretability methods aim to explore key features learned from the input or to identify input features that have a significant impact on the predictions, employing techniques such as convolutional kernel visualization to identify learned sequence motifs, gradient-based methods to quantify feature importance, and perturbation-based analyses to assess the functional impact of input modifications. In contrast, model interpretability approaches, enhance transparency through architectural designs, including attention mechanisms that reveal the relationships between features and biologically inspired transparent models that explicitly align network components with known biological entities. The methods are systematically summarized in Table 1.  The introduction of each method is as follows.

2.1 Input Interpretability

Input interpretability aims to reveal key patterns within the input features or to assess the importance of each feature for the prediction of the model. Unlike traditional k-mer-based feature engineering methods that directly reveal feature importance, deep neural networks employ multi-layer nonlinear transformations for automated feature extraction, resulting in opaque decision processes. To address this, there exist many input interpretability methods, which we further categorize into three classes: convolutional kernel visualization, gradient-based methods, and perturbation methods. Figure 2 illustrates the underlying principles of each method.

Convolutional kernel visualization. With the advancement of deep learning technology, convolutional neural networks (CNNs) have been widely applied in genomic research, such as enhancer prediction [24], transcription factor binding site prediction [41], and chromatin accessibility prediction [14]. CNNs can automatically extract features from DNA sequences. Specifically, CNNs slide multiple filters across the DNA sequence in the convolutional layers, obtaining activation values for each position relative to each filter. These activation values are then used as features and passed to fully connected layers for prediction. As shown in Figure 2(a), the convolutional kernel weights in the first layer are first converted to a position frequency matrix (PFM), followed by log-scaling to produce a standard position weight matrix (PWM) [1, 28], which captures a short sequence motif. Therefore, visualizing these filters can help us understand which sequence patterns the model has extracted that are crucial for prediction. A potential issue here is the constraints in weight learning: unconstrained or inappropriate constraints during weight learning may lead to scaling problems, causing the importance of certain sequences to be disproportionately emphasized or underestimated. In most studies, a commonly used visualization strategy involves computing the activation values of multiple sequences for each trained filter, indicating the degree of match between each sequence and the filter. The sequences with the highest activation values are then selected to generate a corresponding PWM, representing the motif learned by the filter. Then we can obtain motifs learned by multiple filters and match them with existing motif databases to verify whether the model has captured biologically relevant and important features. For example, DeepEnhancer [24] uses a CNN to predict enhancers from DNA sequences. Its first convolutional layer applies 128 filters of length 8 to extract features. By visualizing these filters, the model identifies motifs and matches them with the known transcription factor binding site database JASPAR [23], successfully recognizing biologically relevant motifs. Furthermore, Basset [14] recovered a large number of known DNA-binding protein motifs by analyzing the parameters of the 300 convolutional filters in the convolutional layer.

Refer to caption
Figure 2: Input Interpretability. a. Convolutional kernel visualiza- tion. A single convolutional filter acts as a position weight matrix to scan the input sequence to quantify the binding bias of each position nucleotide in the DNA sequence. b. Perturbation-based methods. Systematic mutagenesis of individual nucleotides or sequence seg- ments enables quantitative evaluation of node-specific importance for model predictions. c. The gradient of the loss function computed by backpropagation highlights the contribution of each input feature.

Gradient-based methods. Gradient-based importance analysis methods achieve global quantitative evaluation of feature importance by computing the partial derivatives of model outputs relative to input features. As shown in Figure 2(c), this approach utilizes backpropagation to compute the gradients of the loss function with respect to sequence features, where gradient magnitudes directly measure the contribution of individual nucleotide in the DNA sequence. For instance, in genomic analyses, positive/negative gradient peaks correspond to enhancer and silencer regions, respectively [15]. However, gradient-based methods are susceptible to vanishing gradients or gradient saturation. In deep neural networks or when using saturating activation functions, excessively small gradients may introduce bias in feature importance estimation. This limitation can be mitigated by integrated gradients, which correct importance scores through path integration [33]. Moreover, an alternative approach, DeepLIFT [31], provides local explanations by comparing the prediction differences between test instances and reference sequences (e.g., background nucleotide frequencies), thereby avoiding the saturation effects of gradient methods. DeepLIFT has been employed to identify critical nucleotides in splice sites [44]. The application of DeepLIFT algorithm successfully identified expression-predictive motifs (EPMs) in both 5’UTR and 3’UTR regions, which exhibit distinct sequence-specific positional preferences [29]. Notably, the attribution rules of DeePLIFT match Shapley values, providing a robust feature importance framework [21]. Both integrated gradients and DeepLIFT require a predefined reference baseline, and the choice of this baseline can significantly influence the accuracy of attribution results. However, there is currently no consensus or established guideline on how to select an appropriate reference.

Table 1: Interpretable Methods and Models Available for Deep Learning
Tool Name and URL Year Strategy Model Application
Puffin [8] (Science)
https://github.com/jzhoulab/puffin
2024 Convolutional kernel visualization U-net Regulatory Genomics
INTERACT [43] (Science Advances)
https://zenodo.org/records/10955827
2025
Convolutional kernel visualization
Attention Mechanism
CNN, Transformer
Regulatory Genomics
Basset [14] (Genome research)
http://www.github.com/davek44/Basset
2016 Convolutional kernel visualization CNN Regulatory Genomics
EpiVerse [19] (Nature Communications)
https://github.com/jhhung/EpiVerse
2025
Convolutional kernel visualization
Perturbation-based methods
Gradient-based methods
CNN, Transformer 3D Genomics
Orca [42] (Nature genetics)
https://github.com/jzhoulab/orca
2022 Perturbation-based methods CNN 3D Genomics
C.origama [34] (Nature biotechnology)
https://github.com/tanjimin/C.Origami
2023
Perturbation-based methods
Gradient-based methods
Attention mechanism
CNN, Transformer
3D Genomics
CLMLA [7] (Science Advances)
https://github.com/PayamDiba/CIMLA
2025 Perturbation-based methods Random Forest Regulatory Genomics
Deepliver [5] (Nature Cell Biology)
https://zenodo.org/record/8139953
2024
Perturbation-based methods
Gradient-based methods
CNN Regulatory Genomics
Geneformer [35] (Nature)
https://huggingface.co/ctheodoris/Geneformer
2023
Perturbation-based methods
Attention mechanism
Transformer Regulatory Genomics
DeepCRE [29] (Nature Communications)
https://github.com/NAMlab/DeepCRE
2024
Gradient-based methods
CNN
Regulatory Genomics
GET [10] (Nature)
https://github.com/GET-Foundation
2025 Gradient-based methods Transformer Regulatory Genomics
Enformer [2] (Nature methods)
https://tfhub.dev/deepmind/enformer/1
2021
Attention mechanism
Perturbation-based methods
Gradient-based methods
Transformer Regulatory Genomics
DCell [22] (Nature methods)
http://d-cell.ucsd.edu/
2018
Transparent models
Gradient-based methods
Perturbation-based methods
Neural Network 3D Genomics
Getnet [36] (Communications Biology)
https://github.com/arnovanhilten/GenNet/
2021 Transparent models Neural Network Regulatory Genomics

Perturbation-based methods. Perturbation-based methods infer the importance of features by manipulating the input features and observing the changes in the output of the model. Perturbation-based interpretability methods were first introduced in the field of computer vision, where specific regions of an image are masked to observe changes in predictions, thereby assessing the importance of those regions [38]. This intuitive approach aligns well with human reasoning and has since been widely adopted in other domains. Similarly to image perturbation, one can consider altering certain specific segments in biological gene sequences to determine the importance of the features corresponding to these segments. For example, we can perturb parts of the sequence and observe changes in the output, as seen in Figure 2(b). The discrepancy between the predictions for these new sequences and the original sequence is quantified as the attribution score, serving as a critical metric for evaluating the functional significance of individual nucleotides in the context of the trained model. It is also feasible to mutate a specific nucleotide within the sequence into each of the other three nucleotides and observe the resultant changes in the output. By repeating this operation for all nucleotides, a matrix of dimensions 4×L4\times L can be obtained, which is commonly referred to as the attribution map [32]. iMAP [20] is used to knock out genes for the discovery of therapeutic targets for diseases. The perturbation model based on β\beta-VAES [4] conducts perturbation analysis on single-cell RNA datasets to predict gene expression changes during gene knockout, toxin response, and embryonic development. A perturbation map [6] was developed to provide a scalable approach for evaluating how specific genetic alterations impact the local, proximal, and distal tumor microenvironment (TME) states. However, many neural networks are trained in a way that resists Dropout [12]. As a result, multiple neurons may be associated with the importance of the same feature, which can lead to the failure of perturbation methods.

For input interpretability, while weight matrices corresponding to convolutional kernels can reflect the importance of sequences, the lack of proper constraints during their learning process may distort the perceived importance of these sequences. Gradient-based methods can evaluate feature importance by computing the partial derivatives of model output with respect to input features, while they may also face the potential issue of gradient vanishing. Perturbation-based methods assess the importance of altered features by introducing minor perturbations to the input and observing the resulting changes in model outputs, yet they may lead to an underestimation of feature importance due to neuronal redundancy.

2.2 Model interpretability

Refer to caption
Figure 3: Model Interpretability. a. Attention Mechanism. Quantify the interactions between different nucleotide positions in the sequence, thereby revealing combinatorial effects on model predictions or biological functions. b. Transparent models. The models are constructed by biological prior knowledge.

Model interpretability aims to facilitate human understanding of internal mechanisms and decision-making processes through architectural design. For example, traditional linear models are inherently interpretable, as we can directly understand the importance of each feature through its assigned weight. In contrast, deep neural networks typically consist of multiple hidden layers and numerous parameters, making it difficult to comprehend their decision-making processes solely based on their architecture or parameters. Therefore, many researchers are currently exploring model-based interpretability techniques to enhance the credibility of the model and uncover new biological insights. These methods can be mainly categorized into two types: attention mechanisms and transparent models based on biological prior knowledge. The principles of the various methods are illustrated in Figure 3.

Attention mechanisms. The attention mechanism is a widely adopted technique in neural networks. A representative example is the Transformer model [37], which leverages self-attention to capture dependencies between tokens within a sequence. The attention weights inherently provide a form of interpretability [27], offering insight into which input tokens the model focuses on when making predictions and revealing interactions between them. The visualization results are shown in Figure 3(a). EpiBERT [13] introduces a masking technique for genomic loci and enhancer regions, extracts query and key matrices from each attention layer, averages the resulting weights across all layers and attention heads, and visualizes these weights to indicate the degree of correlation between different positions. Enformer [2] extracts the query row at the transcription start site (TSS), where the keys denote different spatial positions and the attention values reflect the focus of the model on these positions during the prediction of TSS. In addition, C.Origami [34] exploits attention scores from Transformer models to identify cis-regulatory elements that play critical roles in 3D chromatin architecture. However, in DNA sequences, sequences at multiple positions may be associated with the same biological feature, which is mathematically termed multicollinearity, and this characteristic may lead to instability in the weights of attention mechanism matrices.

Transparent models based on biological prior knowledge. Unlike traditional neural networks, where neurons in the hidden layer typically lack clear biological meaning, transparent neural network models are explicitly designed so that their hidden nodes correspond directly to specific biological units at a fine-grained level, thereby significantly enhancing interpretability [22, 9, 36]. To construct models with intrinsically interpretable units, it is necessary to incorporate prior knowledge into the network architecture design. For instance, by mapping bottom-layer nodes to molecular entities (e.g., genes or proteins) and deeper-layer nodes to functional modules (e.g., metabolic pathways or organelles), a hierarchical biological representation system can be established, as illustrated in Figure 3(b). Deep neural networks build higher-order biological representations through progressive integration of lower-level features, for example, the second layer may capture cooperative interactions between transcription factor binding motifs, while deeper layers can map to complex systems such as complete biological pathways [28]. As a breakthrough in biological applications, DCell [22] pioneered a modeling approach that combines both neural network computational power and model transparency. The input layer of DCell directly corresponds to gene nodes, while its second layer constructs functional group nodes based on the hierarchical structure of Gene Ontology, establishing an association model between genotype and yeast growth rate. The GenNet [36] model utilizes genetic variants (SNPs) as input layer nodes and maps them to gene nodes in the second layer through NCBI RefSeq gene annotations, thereby constructing predictive association models between genotypes and complex phenotypes (e.g., schizophrenia, hair color). Through transparent variable construction based on multi-omics data, P-NET [9] has identified MDM4 as a biomarker in metastatic prostate cancer. Although transparent models offer significant interpretability advantages, their construction relies on domain-specific prior knowledge, which to some extent limits their applicability. Furthermore, the technical approaches used to ensure model transparency may adversely affect predictive performance.

For model interpretability, the attention mechanism reveals the importance of input tokens and reflects the correlations between different tokens. Multicollinearity among input tokens can lead to numerical instability in attention weights. Transparent neural networks can fully incorporate biologically relevant prior knowledge to reveal specific biological functional relationships, yet they may lack universality and cause models to overlook not clearly defined relationships.

3 Theoretical analysis

In this section, we present a systematic theoretical analysis of the technical limitations inherent in each methodological approach.

3.1 The limitation of input interpretability

Analysis reveals three fundamental limitations in input interpretability approaches for deep neural networks: First, unconstrained weight optimization in convolutional layers provably causes scaling instability, mathematically demonstrating disproportionate feature amplification. Second, the inherent neuronal redundancy from dropout mechanisms obscures node-level importance attribution. Third, the null gradient of ReLU for negative inputs, combined with the multiplicative accumulation of small weights through successive layers, results in progressive gradient decay during backpropagation.

The unconstrained learning of weights. CNNs for DNA sequence prediction extract local features by sliding multiple convolutional kernels across the input sequence. The learned kernels can ultimately be interpreted as PWM, representing the key motifs identified by the model from the input sequences. However, it should be noted that the learning of weights should be regulated by constraints. Otherwise, it is highly likely to lead to scaling issues.

Suppose we study a classification problem where a DNA sequence is input to determine whether each position is related to a certain structure. The input sequences are typically represented as a L×4L\times 4 matrix, where sijs_{ij} represents the entry at position (i, j) and the convolutional kernel is a matrix of size k×4k\times 4. To facilitate the analysis of the problem, the network structure is simplified. Specifically, the fully connected layers are abstracted into a single layer, and the binary cross-entropy loss function is adopted as the loss function. The sigmoid function L=σ(z)L=\sigma(z) is used as the activation function for the last layer of the neural network. The weights of the convolutional kernel are updated through gradient descent. Then in each learning iteration, the weights will be updated in the following manner:

wijwijηLwij\footnotesize w_{ij}\xleftarrow{}w_{ij}-\eta\frac{\partial L}{\partial w_{ij}} (1)

To further analyze the scaling problem of the weights, the partial derivatives will be expanded in accordance with the chain rule. Taking w11w_{11}, which is located at the first row and the first column of the convolutional kernel as an example:

Lw11=Lzzyyw11\footnotesize\frac{\partial L}{\partial w_{11}}=\frac{\partial L}{\partial z}\frac{\partial z}{\partial y}\frac{\partial y}{\partial w_{11}} (2)
Lz\displaystyle\frac{\partial L}{\partial z} =y1σ(z)σ(z)(1y)11σ(z)(σ(z))\displaystyle=-y\frac{1}{\sigma(z)}\sigma^{\prime}(z)-(1-y)\frac{1}{1-\sigma(z)}(-\sigma^{\prime}(z)) (3)
=y(1σ(z))+(1y)σ(z)\displaystyle=-y(1-\sigma(z))+(1-y)\sigma(z)
=σ(z)y\displaystyle=\sigma(z)-y

The sequence feature values y1,y2,,yLy_{1},y_{2},...,y_{L} are obtained by convolutional operation (using y1y_{1} as an example):

y1\displaystyle y_{1} =w11s11+w12s12+w13s13+w14s14\displaystyle=w_{11}s_{11}+w_{12}s_{12}+w_{13}s_{13}+w_{14}s_{14}
+w21s21+w22s22+w23s23+w24s24\displaystyle+w_{21}s_{21}+w_{22}s_{22}+w_{23}s_{23}+w_{24}s_{24}
+\displaystyle+...
+wk1sk1+wk2sk2+wk3sk3+wk4sk4\displaystyle+w_{k1}s_{k1}+w_{k2}s_{k2}+w_{k3}s_{k3}+w_{k4}s_{k4}

The output zz is obtained by multiplying the fully-connected weight matrix with the feature sequence.

z=α1y1+α2y2++αLyL\footnotesize z=\alpha_{1}y_{1}+\alpha_{2}y_{2}+...+\alpha_{L}y_{L} (4)

Furthermore, the update formula for the weights of the convolutional kernel can be specifically written.

Lw11=(σ(z)y)(α1s11+α2s21++αLsL1)\footnotesize\frac{\partial L}{\partial w_{11}}=(\sigma(z)-y)(\alpha_{1}s_{11}+\alpha_{2}s_{21}+...+\alpha_{L}s_{L1}) (5)

Among them, sijs_{ij} \in {0,1}\{0,1\}. If the true label is 1, then σ(z)y<0\sigma(z)-y<0. If the subsequent part is greater than 0, it will cause the weight to increase. At this time, if there is no constraint, there will be a risk that the weight will continue to increase. Eventually, a certain weight will become too large.

Then it is common practice to transform the PWM into a PFM[28] through the softmax operation, and then obtain the standard PWM matrix via a logarithmic transformation.

The softmax operation on the convolutional kernel is shown in the following equation (still taking w11w_{11} as an example):

w11=ew11ew11+ew12+ew13+ew14\footnotesize w^{\prime}_{11}=\frac{e^{w_{11}}}{e^{w_{11}}+e^{w_{12}}+e^{w_{13}}+e^{w_{14}}} (6)

It can be seen that if w11w_{11} is too large while the other weight values are relatively small, it leads to w11ew11ew11+0+0+0=1w^{\prime}_{11}\approx\frac{e^{w_{11}}}{e^{w_{11}}+0+0+0}=1.

And if w11w_{11} is small and the other weight is relatively large, it will lead to w1100+ew12+ew13+ew14=0w^{\prime}_{11}\approx\frac{0}{0+e^{w_{12}}+e^{w_{13}}+e^{w_{14}}}=0.

This will lead to the irrationality of the weights. It will cause the functions of certain bases in the sequence to be overemphasized, while the significance of some other bases is neglected.

For the purpose of ensuring the positive and negative nature of the values, when performing the logarithmic transformation, it is usually carried out by wijlog=log(wij+ϵ)w_{ij_{log}}=log(w^{\prime}_{ij}+\epsilon), which ϵ\epsilon is always determined according to the condition to satisfy the stability of the values. Due to ϵ\epsilon being always small, when wijw^{\prime}_{ij} approaches 0, the value after logarithmic transformation will tend to negative infinity. When wijw^{\prime}_{ij} approaches 1, the value after the logarithmic transformation will tend to 0. Both scenarios will lead to the distortion of wijlogw_{ijlog}.

This may lead to the effect of the bases in certain parts of the input DNA sequence being overly amplified, while the roles of the bases at other positions are ignored. This process theoretically analyzes why unconstrained learning can lead to scaling issues, thus rendering the interpretation of the DNA sequence lacking in reliability.

The redundancy of neurons. Perturbation-based methods can evaluate the contribution of individual neurons by selectively removing them and observing the resulting changes in model predictions. However, deep neural networks are usually trained in a way that resists Dropout. The mechanism of Dropout is to deactivate some neurons during each training session. This ensures that certain specific features in the input DNA sequence do not rely on a fixed neuron, but can be captured by multiple neurons. More specifically, it is rjlBernouli(p)r_{j}^{l}\sim Bernouli(p), which means that the jj-th neuron in the ll-th layer has the probability of pp being retained during training process.

Suppose that there are two different neurons, NiN_{i} and NjN_{j}, in the first convolutional layer, yiy_{i} and yjy_{j} are the final contributions of neuron ii and neuron jj in the entire neural network.

yi\displaystyle y_{i} =f(kWikXk+bi)\displaystyle=f(\sum_{k}W_{ik}X_{k}+b_{i})
yj\displaystyle y_{j} =f(kWjkXk+bj)\displaystyle=f(\sum_{k}W_{jk}X_{k}+b_{j})

In standard CNNs, adjacent layer neurons are tightly coupled through weight updates, enabling collaborative feature learning. Dropout randomly disconnects the interneuron pathways during training, forcing independent feature acquisition. This causes redundant node formation: multiple subnetworks repeatedly learn identical DNA sequence features, meaning minimal output changes when removing specific neurons do not necessarily indicate low predictive importance of these nodes.

Since a neural network contains numerous neurons, as well as parameters such as weights and biases, it can be regarded as a high-dimensional manifold \mathcal{M}. The number of neurons in each neural network as a whole, the initial weights and other factors together constitute an abstract parameter θ\theta. From this perspective, the neural network training process is to select a minimum point of the loss function L(θ)L(\theta) on \mathcal{M}. Suppose that the overall parameter settings under the initial conditions are θ\theta. At this time, for any input xx, the output of the model can be expressed as fθ(x)f_{\theta}(x). NiN_{i} and NjN_{j} can be regarded as two distinct minor parameters within the parameter space θ\theta. Removing such a neuron is equivalent to perturbing in a certain parameter direction. If these two neurons are redundant with respect to each other, that is, there exist parameter directions vi,vj𝕕v_{i},v_{j}\in\mathbb{R^{d}} such that for any input xx:

fθ+ϵvi(x)fθ+ϵvj(x)+O(ϵ2)\footnotesize f_{\theta+\epsilon v_{i}}(x)\approx f_{\theta+\epsilon v_{j}}(x)+O(\epsilon^{2}) (7)

This equation indicates that the output results after small perturbations in two directions are approximately equal, where O(ϵ2)O(\epsilon^{2}) denotes an extremely small error related to ϵ\epsilon.

And these two parameter directions can be regarded as symmetric, and the corresponding covariant derivatives commute.

vivjfθ(x)=vjvifθ(x)\footnotesize\nabla v_{i}\nabla v_{j}f_{\theta}(x)=\nabla v_{j}\nabla v_{i}f_{\theta}(x) (8)

Then the components of the Riemann curvature tensor \mathcal{R} of the parameter space in the directions of v1v_{1} and v2v_{2} are:

R(vi,vj,vi,vj)=<vivjfθvjvifθ,fθ>0\footnotesize R(v_{i},v_{j},v_{i},v_{j})=<\nabla v_{i}\nabla v_{j}f_{\theta}-\nabla v_{j}\nabla v_{i}f_{\theta},f_{\theta}>\approx 0 (9)

It represents that the manifold is locally flat in the two parameter directions represented by the removal of redundant nodes. Assume that NiN_{i} is removed. Then it can be regarded as a perturbation in the direction of viv_{i} in the parameter space, with θ=θαvi\theta^{\prime}=\theta-\alpha v_{i}. Subsequently, the variations in the output can be calculated:

Δf\displaystyle\Delta f =fθ(x)fθ(x)\displaystyle=f_{\theta}(x)-f_{\theta^{\prime}}(x)
=fθ(x)fθαv1(x)\displaystyle=f_{\theta}(x)-f_{\theta-\alpha v_{1}}(x)
=θfθ(x)αv112(αv1)Tθ2fθ(x)(αv1)+O(α3)\displaystyle=\nabla_{\theta}f_{\theta}(x)\cdot\alpha v_{1}-\frac{1}{2}(-\alpha v_{1})^{T}\nabla_{\theta}^{2}f_{\theta}(x)(-\alpha v_{1})+O(\alpha^{3})

Due to the large number and complexity of parameters in the entire neural network, the removal of a single neuron can be regarded as a minor change in a certain direction, and α\alpha is an infinitesimal quantity. At the same time, since the manifold \mathbb{R} is locally flat in the subspace spanned by v1v_{1} and v2v_{2}, v1fθ(x)\nabla v_{1}f_{\theta}(x) and v12fθ(x)\nabla v_{1}^{2}f_{\theta}(x) approach 0.

In summary, we obtain that Δf\Delta f approaches 0. That is, when redundancy occurs in neurons due to the Dropout mechanism, the output of the model does not change significantly after removing a single neuron, which makes it hard to access the importance of the node.

Vanishing gradient in DNA sequences. Vanishing gradient is a common issue in gradient-based model interpretation methods, and is closely tied to the choice of activation function. For instance, the ReLU activation function outputs zero for negative inputs while propagating positive inputs unchanged. However, its constant unity gradient for positive inputs may still lead to vanishing gradients during backpropagation in deep networks. Consider a single-layer neural network with the output given by:

Y=ReLU(WX+b)\footnotesize Y=\text{ReLU}(W\cdot X+b) (10)

Here, WW represents the weights matrix, XX is the input, and bb is the bias term. The normal input and output relationship can be expressed as Y=ReLU(WX+b)=WX+bY=\text{ReLU}(W\cdot X+b)=W\cdot X+b. The gradient of the output YY with respect to the input XX can be calculated as:

YX=ReLU(WX+b)X=ReLU(WX+b)W\footnotesize\frac{\partial Y}{\partial X}=\frac{\partial\text{ReLU}(W\cdot X+b)}{\partial X}=\text{ReLU}^{\prime}(W\cdot X+b)\cdot W (11)

According to the derivative definition of ReLU, if WX+b>0W\cdot X+b>0, then ReLU(WX+b)=1\text{ReLU}^{\prime}(W\cdot X+b)=1, and the gradient is WW. Conversely, if WX+b0W\cdot X+b\leq 0, then ReLU(WX+b)=0\text{ReLU}^{\prime}(W\cdot X+b)=0, and the gradient is 0. Extending this to a multi-layer neural network, assume output of each layer passes through a ReLU activation function, with the output of ll-th layer given by:

Z(l)=ReLU(W(l)Z(l1)+b(l))\footnotesize Z^{(l)}=\text{ReLU}(W^{(l)}\cdot Z^{(l-1)}+b^{(l)}) (12)

Here, Z(0)=XZ^{(0)}=X is the input, and Z(L)=YZ^{(L)}=Y is the output. To calculate the gradient of the output YY with respect to the input XX, the chain rule must be applied:

YX=Z(L)Z(L1)Z(L1)Z(L2)Z(1)X\footnotesize\frac{\partial Y}{\partial X}=\frac{\partial Z^{(L)}}{\partial Z^{(L-1)}}\cdot\frac{\partial Z^{(L-1)}}{\partial Z^{(L-2)}}\cdots\frac{\partial Z^{(1)}}{\partial X} (13)

The gradient of each layer is:

Z(l)Z(l1)=ReLU(W(l)Z(l1)+b(l))W(l)\footnotesize\frac{\partial Z^{(l)}}{\partial Z^{(l-1)}}=\text{ReLU}^{\prime}(W^{(l)}\cdot Z^{(l-1)}+b^{(l)})\cdot W^{(l)} (14)

If the input to a layer W(l)Z(l1)+b(l)0W^{(l)}\cdot Z^{(l-1)}+b^{(l)}\leq 0, then ReLU(W(l)Z(l1)+b(l))=0\text{ReLU}^{\prime}(W^{(l)}\cdot Z^{(l-1)}+b^{(l)})=0, and thus, the gradient for that layer is 0. This means that the gradient at this layer vanishes and cannot continue to propagate backward. Additionally, the gradient is also influenced by the weight matrices, as shown by:

YX=Z(L)Z(L1)Z(L1)Z(L2)Z(1)X=W(L)W(L1)W(1)\frac{\partial Y}{\partial X}=\frac{\partial Z^{(L)}}{\partial Z^{(L-1)}}\cdot\frac{\partial Z^{(L-1)}}{\partial Z^{(L-2)}}\cdots\frac{\partial Z^{(1)}}{\partial X}=W^{(L)}\cdot W^{(L-1)}\cdots W^{(1)}

When W(1),W(L1),,W(L)W^{(1)},W^{(L-1)},\ldots,W^{(L)} are very small, the value of the gradient will also be very small, leading to vanishing gradient.

3.2 The limitation of model interpretability

This section systematically demonstrates the inherent limitations of model interpretability approaches in DNA sequence analysis. First, attention mechanisms intrinsically exhibit numerical instability, which can be mathematically proven to induce matrix ill-conditioning under multicollinearity. Second, biologically-constrained transparent models may yield higher loss values compared to their unconstrained counterparts.

Instability of in the estimation of the weights. The computation of attention weights typically involves two vectors, namely the query vector and the key vector. The attention score is calculated by taking the dot product of the query matrix Q=[q1,q2,,qn]{Q}=[q_{1},q_{2},...,q_{n}] and the key matrix K=[k1,k2,,kn]{K}=[k_{1},k_{2},...,k_{n}]:

S=QKT\footnotesize S=QK^{T} (15)

Since the input features often exhibit multicollinearity, some of the vectors in the key matrix are linearly correlated. That is, there exists an index ii such that ki=jiαjkj+βk_{i}=\sum_{j\neq i}\alpha_{j}k_{j}+\beta.

Consider the covariance matrix CC of KK, and C=KTKC=K^{T}K, due to the multicollinearity, the rank of the matrix KK will satisfy r(K)=r<nr(K)=r<n, then we have:

r(C)=r(KTK)r(K)<n\footnotesize r(C)=r(K^{T}K)\leq r(K)<n (16)

The matrix CC also satisfies the condition of partial linear correlation. Therefore, |C|=0|C|=0. Since the matrix CC is a positive semi-definite matrix, all the eigenvalues of CC are greater than or equal to 0. At this time, there will be:

|C|=λ1λ2λn=0\footnotesize|C|=\lambda_{1}\lambda_{2}\cdot\cdot\cdot\lambda_{n}=0 (17)

So there must exist λi\lambda_{i} which satisfies that λi=0\lambda_{i}=0, i.e., λmin=0\lambda_{min}=0.

To demonstrate that multicollinearity leads to numerical instability, we first present the following lemma and theorem.

Lemma 1.

In the sense of the 2-norm, for any matrix AA, if λmax\lambda_{max} is the largest eigenvalue of ATAA^{T}A, then A2=λmax||A||_{2}=\sqrt{\lambda_{max}}.

Proof.

Arbitrarily select a matrix AA and any unit vector xx that can be transformed by AA. According to the definition of the norm and the unit vector, we have:

A2=supx0Axx=supx0Ax\footnotesize||A||_{2}=sup_{||x||\neq 0}\frac{||Ax||}{||x||}=sup_{||x||\neq 0}||Ax|| (18)

Then write Ax||Ax|| in another form:

Ax=(Ax)T(Ax)=xTATAx\footnotesize||Ax||=\sqrt{(Ax)^{T}(Ax)}=\sqrt{x^{T}A^{T}Ax} (19)

Due to xTATAx0x^{T}A^{T}Ax\geq 0, ATAA^{T}A is a positive semi-definite matrix. Suppose that its nn eigenvalues satisfy λ1λ2λn0\lambda_{1}\geq\lambda_{2}\geq...\geq\lambda_{n}\geq 0, the corresponding eigenvectors α1,α2,,αn\alpha_{1},\alpha_{2},...,\alpha_{n} form an orthonormal basis of n\mathbb{R}^{n}. Let x=i=1nkiαix=\sum_{i=1}^{n}k_{i}\alpha_{i}, since x=1||x||=1, so i=1nki2=1\sum_{i=1}^{n}k_{i}^{2}=1. Subsequently, we can derive the following equation:

ATAx=ATA(i=1nkiαi)=i=1nkiATAαi=i=1nλikiαi\footnotesize A^{T}Ax=A^{T}A(\sum_{i=1}^{n}k_{i}\alpha_{i})=\sum_{i=1}^{n}k_{i}A^{T}A\alpha_{i}=\sum_{i=1}^{n}\lambda_{i}k_{i}\alpha_{i} (20)
xTATAx=i=1nλiki2λ1\footnotesize x^{T}A^{T}Ax=\sum_{i=1}^{n}\lambda_{i}k_{i}^{2}\leq\lambda_{1} (21)

λ1\lambda_{1} is the largest eigenvalue of ATAA^{T}A, which is marked as λmax\lambda_{max}. Then, (Ax)T(Ax)λmax\sqrt{(Ax)^{T}(Ax)}\leq\sqrt{\lambda_{max}}.

Up to this point, we have obtained A2=λmax||A||_{2}=\sqrt{\lambda_{max}}.

Theorem 2.

Multicolinearity leads to the condition number of the matrix tending to infinity.

Proof.

According to Lemma 1, we have A2=λmax||A||_{2}=\sqrt{\lambda_{max}} and A12=1λmin||A^{-1}||_{2}=\frac{1}{\sqrt{\lambda_{min}}}.

k(A)=A2A12=λmaxλmin\footnotesize k(A)=||A||_{2}\cdot||A^{-1}||_{2}=\frac{\sqrt{\lambda_{max}}}{\sqrt{\lambda_{min}}}\rightarrow\infty (22)

This means that multicollinearity will lead to an excessively large condition number of the matrix, which in turn poses a potential risk of numerical instability [40].

Constructing transparent models with prior knowledge. Transparent models constructed based on prior knowledge have attracted widespread attention due to their computational efficiency and interpretability. However, transparent models also have their limitations. Biological constraints within transparent models may diminish their predictive performance. For example, Hard-coding TF-TF interactions may underperform unconstrained models. (although the use of milder regularization techniques can mitigate this decrease to some extent). The following serves as an illustration of this. Suppose that the DNA samples input into two models are identical, with all other confounding factors excluded. The loss function for the hard-coded transparent model is defined as follows:

Lhard=i=1Nmilog(pihard)+λWpriorWref_prior2+γWadapt2\footnotesize L_{\text{hard}}=-\sum_{i=1}^{N}m_{i}\log(p_{i}^{\text{hard}})+\lambda\|W_{\text{prior}}-W_{\text{ref\_prior}}\|_{2}+\gamma\|W_{\text{adapt}}\|_{2} (23)

There, NN denotes the number of input samples, mim_{i} is the true label of the ii-th sample, and pihardp_{i}^{\text{hard}} represents the predicted probability of the ii-th sample by the hard-coded model. This prediction is subject to the constraint pihard=f(Wprior,Wadapt,mi)p_{i}^{\text{hard}}=f(W_{\text{prior}},W_{\text{adapt}},m_{i}). The regularization coefficient λ\lambda is associated with the prior constraint term, while γ\gamma corresponds to the regularization coefficient for the adjustable parameters in the hard-coded model, which usually has a smaller value. WpriorW_{\text{prior}} signifies the fixed weights in the hard-coded model, Wref_priorW_{\text{ref\_prior}} is the preset baseline weight reference value, and WadaptW_{\text{adapt}} is the weight in the hard-coded model that can be freely adjusted. The loss function of the hard-coded model is primarily determined by the first term, which is the cross-entropy loss. The loss function for the unconstrained model is defined as follows:

Lfree=i=1Nmilog(pifree)+βWfree2\footnotesize L_{\text{free}}=-\sum_{i=1}^{N}m_{i}\log(p_{i}^{\text{free}})+\beta\|W_{\text{free}}\|_{2} (24)

In comparison, the predicted probability pifree=f(Wfree,mi)p_{i}^{\text{free}}=f(W_{\text{free}},m_{i}) of the unconstrained model is determined by the free parameters WfreeW_{\text{free}}. Here, β\beta is the regularization coefficient for the unconstrained model (with a smaller value). Similarly, the loss function of the unconstrained model is primarily determined by the first term, which is the cross-entropy loss. Due to the limitations imposed by hard coding, it holds that pihardpifreep_{i}^{\text{hard}}\leq p_{i}^{\text{free}}. To demonstrate that the performance of the hard-coded model may potentially decrease compared to the unconstrained model, it is only necessary to prove that the difference between the loss functions of the hard-coded model and the unconstrained model is not less than zero.

LhardLfree\displaystyle L_{\text{hard}}-L_{\text{free}} =i=1Nmi(logpifreepihard)+λWpriorWref_prior2\displaystyle=\sum_{i=1}^{N}m_{i}\left(\log\frac{p_{i}^{\text{free}}}{p_{i}^{\text{hard}}}\right)+\lambda\|W_{\text{prior}}-W_{\text{ref\_prior}}\|_{2}
+γWadapt2βWfree2\displaystyle\quad+\gamma\|W_{\text{adapt}}\|_{2}-\beta\|W_{\text{free}}\|_{2}

Where i=1Nmi(logpifreepihard)0\sum_{i=1}^{N}m_{i}\left(\log\frac{p_{i}^{\text{free}}}{p_{i}^{\text{hard}}}\right)\geq 0, λWpriorWref_prior20\lambda\|W_{\text{prior}}-W_{\text{ref\_prior}}\|_{2}\geq 0, and γWadapt2βWfree20\gamma\|W_{\text{adapt}}\|_{2}-\beta\|W_{\text{free}}\|_{2}\approx 0. From this, it can be concluded that LhardLfree0L_{\text{hard}}-L_{\text{free}}\geq 0, which proves that the hard-coded model may perform worse than the unconstrained model.

4 Future Directions

Although we have theoretically analyzed the problems existing in various interpretable methods, how to optimize and improve these methods with limitations remains a direction worthy of exploration.

Overall, on the one hand, research demonstrates that existing interpretation approaches, including attention-based visualization techniques, feature importance analysis, and causal reasoning-exhibit significant variations in terms of information presentation formats, analytical granularity, and reliability. Such heterogeneity not only poses substantial challenges for researchers in selecting and applying appropriate methods, but may also potentially compromise the credibility of model-derived decisions to some extent. Consequently, establishing a systematic evaluation framework for explainable methods becomes particularly crucial. Future research should focus on developing comprehensive and objective assessment metrics to construct a unified benchmarking standard, thereby effectively enhancing the reliability and practical utility of explainability studies.

On the other hand, we expect that interpretable information will not rely solely on a single method or a certain input information. With the development of AI and the improvement of computing power, multimodal technology is thriving. In the future, multimodal technology can be applied to integrate various forms of input data, such as gene sequence and protein structures. At the same time, by comprehensively considering the information provided by different types of interpretation methods, the understanding and reasoning abilities of the model can be enhanced, providing a more reliable guarantee for genetic predictions.

5 Conclusions

With the extensive application of deep learning in the field of bioinformatics, reliable interpretability has become increasingly crucial for understanding the prediction results. This article provides a systematic overview of interpretability methods for deep learning applied in the field of bioinformatics and then categorizes these methods into two primary types: input interpretability and model interpretability, further detailing various interpretive approaches within each category. Additionally, we theoretically analyze the limitations of these interpretive methods and provide guidance for future research on interpretability in the field of genetics.

References

  • Alipanahi et al. [2015] B. Alipanahi, A. Delong, M. T. Weirauch, and B. J. Frey. Predicting the sequence specificities of dna-and rna-binding proteins by deep learning. Nature biotechnology, 33(8):831–838, 2015.
  • Avsec et al. [2021] Ž. Avsec, V. Agarwal, D. Visentin, J. R. Ledsam, A. Grabska-Barwinska, K. R. Taylor, Y. Assael, J. Jumper, P. Kohli, and D. R. Kelley. Effective gene expression prediction from sequence by integrating long-range interactions. Nature methods, 18(10):1196–1203, 2021.
  • Azodi et al. [2020] C. B. Azodi, J. Tang, and S.-H. Shiu. Opening the black box: interpretable machine learning for geneticists. Trends in genetics, 36(6):442–455, 2020.
  • [4] A. Bjerregaard, V. Das, and A. Krogh. Interpretable single-cell perturbations from decoder gradients. In ICLR 2025 Workshop on Machine Learning for Genomics Explorations.
  • Bravo Gonzalez-Blas et al. [2024] C. Bravo Gonzalez-Blas, I. Matetovici, H. Hillen, I. I. Taskiran, R. Vandepoel, V. Christiaens, L. Sansores-Garcia, E. Verboven, G. Hulselmans, S. Poovathingal, et al. Single-cell spatial multi-omics and deep learning dissect enhancer-driven gene regulatory networks in liver zonation. Nature Cell Biology, 26(1):153–167, 2024.
  • Dhainaut et al. [2022] M. Dhainaut, S. A. Rose, G. Akturk, A. Wroblewska, S. R. Nielsen, E. S. Park, M. Buckup, V. Roudko, L. Pia, R. Sweeney, et al. Spatial crispr genomics identifies regulators of the tumor microenvironment. Cell, 185(7):1223–1239, 2022.
  • Dibaeinia et al. [2025] P. Dibaeinia, A. Ojha, and S. Sinha. Interpretable ai for inference of causal molecular relationships from omics data. Science Advances, 11(7):eadk0837, 2025.
  • Dudnyk et al. [2024] K. Dudnyk, D. Cai, C. Shi, J. Xu, and J. Zhou. Sequence basis of transcription initiation in the human genome. Science, 384(6694):eadj0116, 2024.
  • Elmarakeby et al. [2021] H. A. Elmarakeby, J. Hwang, R. Arafeh, J. Crowdis, S. Gang, D. Liu, S. H. AlDubayan, K. Salari, S. Kregel, C. Richter, et al. Biologically informed deep neural network for prostate cancer discovery. Nature, 598(7880):348–352, 2021.
  • Fu et al. [2025] X. Fu, S. Mo, A. Buendia, A. P. Laurent, A. Shao, M. d. M. Alvarez-Torres, T. Yu, J. Tan, J. Su, R. Sagatelian, et al. A foundation model of transcription across human cell types. Nature, pages 1–9, 2025.
  • Fudenberg et al. [2020] G. Fudenberg, D. R. Kelley, and K. S. Pollard. Predicting 3d genome folding from dna sequence with akita. Nature methods, 17(11):1111–1117, 2020.
  • Hinton et al. [2012] G. E. Hinton, N. Srivastava, A. Krizhevsky, I. Sutskever, and R. R. Salakhutdinov. Improving neural networks by preventing co-adaptation of feature detectors. arXiv preprint arXiv:1207.0580, 2012.
  • Javed et al. [2025] N. Javed, T. Weingarten, A. Sehanobish, A. Roberts, A. Dubey, K. Choromanski, and B. E. Bernstein. A multi-modal transformer for cell type-agnostic regulatory predictions. Cell Genomics, 2025.
  • Kelley et al. [2016] D. R. Kelley, J. Snoek, and J. L. Rinn. Basset: learning the regulatory code of the accessible genome with deep convolutional neural networks. Genome research, 26(7):990–999, 2016.
  • Kelley et al. [2018] D. R. Kelley, Y. A. Reshef, M. Bileschi, D. Belanger, C. Y. McLean, and J. Snoek. Sequential regulatory activity prediction across chromosomes with convolutional neural networks. Genome research, 28(5):739–750, 2018.
  • Klemm et al. [2019] S. L. Klemm, Z. Shipony, and W. J. Greenleaf. Chromatin accessibility and the regulatory epigenome. Nature Reviews Genetics, 20(4):207–220, 2019.
  • Koo and Ploenzke [2020] P. K. Koo and M. Ploenzke. Deep learning for inferring transcription factor binding sites. Current opinion in systems biology, 19:16–23, 2020.
  • Lieberman-Aiden et al. [2009] E. Lieberman-Aiden, N. L. Van Berkum, L. Williams, M. Imakaev, T. Ragoczy, A. Telling, I. Amit, B. R. Lajoie, P. J. Sabo, M. O. Dorschner, et al. Comprehensive mapping of long-range interactions reveals folding principles of the human genome. science, 326(5950):289–293, 2009.
  • Lin et al. [2025] M.-Y. Lin, Y.-C. Lo, and J.-H. Hung. Unveiling chromatin dynamics with virtual epigenome. Nature Communications, 16(1):3491, 2025.
  • Liu et al. [2022] B. Liu, Z. Jing, X. Zhang, Y. Chen, S. Mao, R. Kaundal, Y. Zou, G. Wei, Y. Zang, X. Wang, et al. Large-scale multiplexed mosaic crispr perturbation in the whole organism. Cell, 185(16):3008–3024, 2022.
  • Lundberg and Lee [2017] S. M. Lundberg and S.-I. Lee. A unified approach to interpreting model predictions. Advances in neural information processing systems, 30, 2017.
  • Ma et al. [2018] J. Ma, M. K. Yu, S. Fong, K. Ono, E. Sage, B. Demchak, R. Sharan, and T. Ideker. Using deep learning to model the hierarchical structure and function of a cell. Nature methods, 15(4):290–298, 2018.
  • Mathelier et al. [2016] A. Mathelier, O. Fornes, D. J. Arenillas, C.-y. Chen, G. Denay, J. Lee, W. Shi, C. Shyr, G. Tan, R. Worsley-Hunt, et al. Jaspar 2016: a major expansion and update of the open-access database of transcription factor binding profiles. Nucleic acids research, 44(D1):D110–D115, 2016.
  • Min et al. [2017] X. Min, W. Zeng, S. Chen, N. Chen, T. Chen, and R. Jiang. Predicting enhancers with deep convolutional neural networks. BMC bioinformatics, 18:35–46, 2017.
  • Molnar et al. [2020] C. Molnar, G. Casalicchio, and B. Bischl. Interpretable machine learning–a brief history, state-of-the-art and challenges. In Joint European conference on machine learning and knowledge discovery in databases, pages 417–431. Springer, 2020.
  • Mumbach et al. [2016] M. R. Mumbach, A. J. Rubin, R. A. Flynn, C. Dai, P. A. Khavari, W. J. Greenleaf, and H. Y. Chang. Hichip: efficient and sensitive analysis of protein-directed genome architecture. Nature methods, 13(11):919–922, 2016.
  • Naim and Asher [2024] O. Naim and N. Asher. On explaining with attention matrices. In ECAI 2024, pages 1035–1042. IOS Press, 2024.
  • Novakovsky et al. [2023] G. Novakovsky, N. Dexter, M. W. Libbrecht, W. W. Wasserman, and S. Mostafavi. Obtaining genetics insights from deep learning via explainable artificial intelligence. Nature Reviews Genetics, 24(2):125–137, 2023.
  • Peleke et al. [2024] F. F. Peleke, S. M. Zumkeller, M. Gültas, A. Schmitt, and J. Szymański. Deep learning the cis-regulatory code for gene expression in selected model plants. Nature Communications, 15(1):3488, 2024.
  • Schuette et al. [2025] G. Schuette, Z. Lao, and B. Zhang. Chromogen: Diffusion model predicts single-cell chromatin conformations. Science Advances, 11(5):eadr8265, 2025.
  • Shrikumar et al. [2017] A. Shrikumar, P. Greenside, and A. Kundaje. Learning important features through propagating activation differences. In International conference on machine learning, pages 3145–3153. PMlR, 2017.
  • Simonyan et al. [2013] K. Simonyan, A. Vedaldi, and A. Zisserman. Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv preprint arXiv:1312.6034, 2013.
  • Sundararajan et al. [2016] M. Sundararajan, A. Taly, and Q. Yan. Gradients of counterfactuals. arXiv preprint arXiv:1611.02639, 2016.
  • Tan et al. [2023] J. Tan, N. Shenker-Tauris, J. Rodriguez-Hernaez, E. Wang, T. Sakellaropoulos, F. Boccalatte, P. Thandapani, J. Skok, I. Aifantis, D. Fenyö, et al. Cell-type-specific prediction of 3d chromatin organization enables high-throughput in silico genetic screening. Nature biotechnology, 41(8):1140–1150, 2023.
  • Theodoris et al. [2023] C. V. Theodoris, L. Xiao, A. Chopra, M. D. Chaffin, Z. R. Al Sayed, M. C. Hill, H. Mantineo, E. M. Brydon, Z. Zeng, X. S. Liu, et al. Transfer learning enables predictions in network biology. Nature, 618(7965):616–624, 2023.
  • van Hilten et al. [2021] A. van Hilten, S. A. Kushner, M. Kayser, M. A. Ikram, H. H. Adams, C. C. Klaver, W. J. Niessen, and G. V. Roshchupkin. Gennet framework: interpretable deep learning for predicting phenotypes from genetic data. Communications biology, 4(1):1094, 2021.
  • Vaswani et al. [2017] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin. Attention is all you need. Advances in neural information processing systems, 30, 2017.
  • Zeiler and Fergus [2014] M. D. Zeiler and R. Fergus. Visualizing and understanding convolutional networks. In Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part I 13, pages 818–833. Springer, 2014.
  • Zhang et al. [2025] G. Zhang, C. Song, M. Yin, L. Liu, Y. Zhang, Y. Li, J. Zhang, M. Guo, and C. Li. Trapt: a multi-stage fused deep learning framework for predicting transcriptional regulators based on large-scale epigenomic data. Nature Communications, 16(1):3611, 2025.
  • Zhang et al. [2024] Y. Zhang, H. Zhu, Z. Song, Y. Chen, X. Fu, Z. Meng, P. Koniusz, and I. King. Geometric view of soft decorrelation in self-supervised learning. In Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pages 4338–4349, 2024.
  • Zheng et al. [2021] A. Zheng, M. Lamkin, H. Zhao, C. Wu, H. Su, and M. Gymrek. Deep neural networks identify sequence context features predictive of transcription factor binding. Nature machine intelligence, 3(2):172–180, 2021.
  • Zhou [2022] J. Zhou. Sequence-based modeling of three-dimensional genome architecture from kilobase to chromosome scale. Nature genetics, 54(5):725–734, 2022.
  • Zhou et al. [2025] J. Zhou, D. R. Weinberger, and S. Han. Deep learning predicts dna methylation regulatory variants in specific brain cell types and enhances fine mapping for brain disorders. Science Advances, 11(1):eadn1870, 2025.
  • Zuallaert et al. [2018] J. Zuallaert, F. Godin, M. Kim, A. Soete, Y. Saeys, and W. De Neve. Splicerover: interpretable convolutional neural networks for improved splice site prediction. Bioinformatics, 34(24):4180–4188, 2018.