This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Memory Faults in Activation-sparse Quantized Deep Neural Networks: Analysis and Mitigation using Sharpness-aware Training

Akul Malhotra and Sumeet Kumar Gupta This work is supported by the Center for Brain-Inspired Computing (C-BRIC) and Center for Co-Design of Cognitive Systems (CoCoSys), funded by SRC and DARPA. We would also like to acknowledge Prof. Anand Raghunathan (Purdue University) and Dr. Sarada Krithivasan (IBM) for their suggestions. Corresponding Author: Akul Malhotra. Email: malhot23@purdue.edu. Purdue University, West Lafayette, Indiana
Abstract

Improving the hardware efficiency of deep neural network (DNN) accelerators with techniques such as quantization and sparsity enhancement have shown an immense promise. However, their inference accuracy in non-ideal real-world settings (such as in the presence of hardware faults) is yet to be systematically analyzed. In this work, we investigate the impact of memory faults on activation-sparse quantized DNNs (AS QDNNs). We show that a high level of activation sparsity comes at the cost of larger vulnerability to faults, with AS QDNNs exhibiting up to 11.13% lower accuracy than the standard QDNNs. We establish that the degraded accuracy correlates with a sharper minima in the loss landscape for AS QDNNs, which makes them more sensitive to perturbations in the weight values due to faults. Based on this observation, we employ sharpness-aware quantization (SAQ) training to mitigate the impact of memory faults. The AS and standard QDNNs trained with SAQ have up to 19.50% and 15.82% higher inference accuracy, respectively compared to their conventionally trained equivalents. Moreover, we show that SAQ-trained AS QDNNs show higher accuracy in faulty settings than standard QDNNs trained conventionally. Thus, sharpness-aware training can be instrumental in achieving sparsity-related latency benefits without compromising on fault tolerance.

Index Terms:
DNN Accelerators, Activation Sparsity, Flat minima, Fault Tolerance.

I Introduction

The remarkable success of deep neural networks (DNNs) for decision-making and sensory processing tasks has prompted the exploration of DNN edge accelerators [1]. However, the performance benefits of state-of-the-art DNNs come at the cost of large storage and computation requirements, introducing various design challenges, especially for energy- and memory-constrained edge applications. The need to reduce the size and computational complexity of DNNs has led to the emergence of Quantized Deep Neural networks (QDNNs). QDNNs utilize reduced bit precisions for their weights and activations to achieve energy and latency savings associated with storage, computation and communication, while maintaining an accuracy comparable to their full-precision counterparts [2].

Refer to caption
Figure 1: (a) Shows an activation-sparse QDNN (AS QDNN) with memory faults. We show that AS-QDNNs suffer larger accuracy degradation due to faults than their standard counterparts. (b) describes the trade-off between the latency benefits due to enhanced activation sparsity and reduced fault tolerance. To overcome this, we utilize sharpness-aware quantization (SAQ) training, which enhances the AS QDNNs fault tolerance by flattening its weight loss landscape.

Another popular method to enhance DNN hardware efficiency is to sparsify the weights and activations. Various techniques to induce weight sparsity, also referred to as weight pruning, have led to significant resource savings with little or no accuracy loss [3]. Recently, approaches to leverage and increase the activation sparsity have also gained attention to improve inference speed [4] [5]. Such techniques involving weight and activation sparsification can be used in conjunction with quantization to design QDNNs optimized for edge applications.

While sparsified QDNNs offer promising attributes for resource-constrained systems, their deployment in real world settings also needs to consider hardware defects. DNN weights are generally stored on-chip in a memory which experiences various types of faults such as stuck-at and bit-flip faults. These faults corrupt the weight values and degrade system accuracy. Although the impact of faults in conventional DNNs is well understood, the fault tolerance of weight/activation sparse DNNs has been sparsely studied. Some works have explored the impact of faults and non-idealities on weight pruned models [6][7]. However, the understanding of the impact of faults on activation-sparse DNNs (AS QDNNs) is lacking. Moreover, there is a need for techniques that can alleviate the adverse effect of faults on the accuracy of AS QDNNs.

In this paper, we address these critical needs by (a) analyzing the inference accuracy of AS QDNNs in the presence of memory faults and (b) enhancing their fault tolerance by utilizing Sharpness-Aware Quantization (SAQ) [8]. The key contributions of our work are (summarized in Figure 1) :

  • We show that although an increase in activation sparsity can lead to latency benefits, it comes at the expense of reduced fault tolerance and lower inference accuracy of QDNNs. To the best of our knowledge, this is the first work exploring the latency-accuracy trade-off for AS QDNNs in the presence of faults.

  • Using the weight loss landscape visualization method [9], we establish that the higher sensitivity of AS QDNNs to faults is attributed to sharper minima in their loss landscape (compared to the standard QDNNs).

  • Based on the above finding, we employ SAQ to mitigate the impact of faults on system accuracy [8]. We show that the SAQ-training increases the inference accuracy of AS and standard QDNNs by up to 19.50% and 15.82% (compared to conventional training schemes) by reducing the sharpness of the weight loss landscape.

  • We also show that SAQ-trained AS QDNNs have higher inference accuracy than their standard counterparts trained without SAQ. This enables the design of QDNNs which are both activation-sparse and fault tolerant.

II Background and Related Work

II-A Activation sparsity in DNNs

Activation sparsity refers to the prevalence of a large number of zero-valued DNN activations. A common activation function in DNNs is the rectified linear unit (ReLU), which outputs a zero for every negative input, leading to a high activation sparsity. This motivates the design of sparsity-aware DNN hardware, that can skip the computations involving zero-valued activations, leading to a reduction in the inference latency [10]. Unlike weight sparsity, activation sparsity is dynamic in nature, which means that the number and location of the zero values vary from input to input. When utilized properly, the inherently large activation sparsity can be exploited by sparsity-aware DNN accelerators for sizeable latency improvements over standard accelerators [1] [10].

Consequently, algorithmic techniques to enhance the activation sparsity have gained interest in recent times. For example, the work in [5] adds the L1L_{1} norm of the activations (xl,n1||x_{l,n}||_{1}) to the original loss function to incentivize sparsity:

Lreg(x,w)=L0(x,w)+n=1Nl=1Lαlxl,n1L_{reg}(x,w)=L_{0}(x,w)+\sum^{N}_{n=1}\sum^{L}_{l=1}\alpha_{l}||x_{l,n}||_{1} (1)

Here, Lreg(x,w)L_{reg}(x,w) and L0(x,w)L_{0}(x,w) are the new and original loss functions, respectively, LL is the number of layers in the DNN, NN is the batch size and αl\alpha_{l} is the regularization constant per layer. By penalizing the dense activations using L1L1 regularization, the work in [5] exhibits up to 60% increase in sparsity with negligible loss in inference accuracy for image classification. The work in [4] uses a Hoyer sparsity metric based regularizer and a variant of ReLU to boost the activation sparsity of DNNs. It should be noted that these techniques are not equivalent to ”pruning” the activations [3]. While pruning permanently sets the parameters to zero for all inputs, the activation sparsification techniques do not remove any of the activations. Rather, they ensure that a low percentage of activations are non-zero for all inputs on an average.

In this paper, we use the L1L1 regularization based approach to train activation-sparse (AS) QDNNs. We will refer to the QDNNs trained without this regularization as standard QDNNs and the ones trained with it as AS QDNNs.

II-B Memory faults in DNN accelerators

Aggressive scaling and exploration of new memory technologies have heightened the importance of studying memory faults. Faults corrupt a percentage of the stored weights, leading to erroneous computations. Two common memory fault models are bit-flip faults and stuck-at (SA) faults. Stuck-at one (SA1) and stuck-at zero (SA0) faults fix the value of a bitcell unalterably to ’1’ and ’0’, respectively. Conversely, bit-flip faults invert the value in the bitcell (from ’0’ to ’1’ or vice versa). The SA fault model is generally used to describe permanent defects such as fabrication defects, whereas the bit-flip fault model is used to describe transient phenomena such as half-select read disturbance and alpha particle strikes.

Understanding the impact of memory faults on DNN accuracy has attracted attention in recent times. Various works have analyzed the fault tolerance of both floating-point DNNs and QDNNs [11]. Some works have also studied the accuracy of pruned networks in the presence of faults [7], circuit non-idealities and variations [6], showing a larger vulnerability of pruned DNNs to faults and other non-idealities than their unpruned counterparts. However, to the best of our knowledge, no work has analyzed the impact of faults on AS QDNNs.

Multiple hardware and algorithmic solutions have been analyzed for mitigating the impact of faults. From the hardware perspective, the work in [12] proposes a selective duplication strategy to replicate the vulnerable portions of the DNNs to enhance fault tolerance. The work in [13] leverages the natural redundancy present in compute-in-memory based ternary bitcells to reduce fault impact. From the algorithmic perspective, the work in [14] uses error-correcting output codes to reduce DNN sensitivity to variations and SA faults. Works like [15] and [16] use weight transformations to reduce the amount of unmasked faults, leading to increased fault tolerance. Fault-aware training/finetuning is also a promising approach but is difficult to implement when the fault data is unavailable [17].

In this work, we extensively investigate the impact of SA and bit-flip faults on the performance of standard and AS QDNNs. Additionally, we introduce SAQ-based training as a strategy for fault tolerance. SAQ can be used alongside hardware techniques and offers protection against multiple types of faults, as discussed subsequently.

II-C Sharpness-Aware Quantization (SAQ)

The sharpness of the minima of the loss landscape that a DNN converges to during training has a key impact on its generalization capability [18]. Both theoretically and empirically, it has been shown that convergence to a flat minima improves generalization. Hence, training schemes such as sharpness-aware minimization (SAM) are gaining attention [19]. SAM simultaneously minimizes the loss value and the sharpness of the weight loss landscape to learn the optima which have uniformly low loss values in their neighbourhood. DNNs trained with SAM have displayed state-of-the-art accuracies for several datasets and models. However, SAM is not as effective in QDNNs as it does not account for weight quantization. Sharpness-aware quantization (SAQ) is a variant of SAM designed for QDNNs [8]. SAQ simultaneously minimizes the loss value and flattens the loss curvature near the minima by minimizing the loss function with adversarially perturbed quantized weights. This leads to better generalization and improved accuracy of QDNNs. In this work, we advance the application of SAQ to train fault-tolerant QDNNs, based on the intuition that QDNNs converged at flatter minima will be less sensitive to the weight perturbations due to faults.

III Impact of Faults on Activation-Sparse QDNNs

III-A Experimental Framework

To explore the latency benefits and the impact of faults on activation-sparse (AS) QDNNs, we use the following methodology. First, two 4-bit standard QDNNs (baselines), viz. LeNet-5 with the FashionMNIST (FMNIST) dataset and ResNet-18 with the CIFAR-10 dataset are trained using the quantization framework in [2]. Then, we train the AS QDNNs by combining the L1L1 activation regularization [5] with the quantization framework in [2]. We focus on these model architectures because of their compact size, making them suitable for resource-constrained edge accelerators. The standard and AS QDNNs are trained to have nearly equal inference accuracies (with 0.5%\leq 0.5\% difference) for a fair comparison.

After training, we evaluate their latencies on sparsity-aware hardware based on the work in [10], to quantify the benefits of enhanced sparsity. Then, to study their fault tolerance, bit-flip faults and stuck-at faults (SA0 and SA1 individually) are randomly and uniformly injected into each model. Monte Carlo based fault injection experiments are performed, and the mean value of the inference accuracy is reported for different fault rates and types of faults (see Fig. 5). We perform experiments for fault rates 0% - 5% and 0% - 3% for the LeNet-5 and ResNet-18 QDNNs, respectively. ResNet-18 QDNNs were analyzed up to a fault rate of 3%, as higher rates caused severe accuracy drops, rendering the QDNNs unusable.

To better understand the differences in fault tolerance, we visualize the weight loss landscape for standard and AS QDNNs. Loss landscape visualization has been used in various works to enhance understanding of how well a model generalizes [18]. To our knowledge, this is the first work to examine the shape of the minima in relation to the QDNN’s fault tolerance. We use the normalization-based visualization technique described in [9] to generate the weight loss landscapes.

III-B Results

III-B1 Latency benefits of Enhanced Activation Sparsity

First, we analyse the activation sparsity increase due to L1L1 activation regularization. Figures 2a and 2b show the percentage of zero activations (averaged over all layers and test inputs) for LeNet-5 and ResNet-18 QDNNs, respectively. We observe an activation sparsity increase from 43% (with standard training) to 84% (with L1L1 activation regularization) for LeNet-5 and from 53% to 75% for ResNet-18 QDNNs. We further analyze the sparsity of AS QDNNs in faulty settings. We observe that even at a 5% (3%) bit flip fault rate, the LeNet-5 (ResNet-18) AS QDNN shows only a \sim4% (\sim2%) reduction in sparsity. This signifies that the gain in activation sparsity is sustained even in the presence of faults.

Refer to caption
Figure 2: The activation sparsities of (a) LeNet 5 and (b) ResNet 18 standard QDNNs and activation sparse (AS) QDNNs in both fault-free and faulty settings. The activation sparsity of LeNet 5 and ResNet 18 AS QDNNs is 95.34% and 41.51% higher than their standard counterparts, and is sustained in faulty environments.
Refer to caption
Figure 3: Latencies of various QDNNs deployed on DNN accelerators

Next, we assess the latency benefits provided by the increased sparsity on a sparsity-aware accelerator based on the Cnvlutin architecture [10]. Cnvlutin skips computations with zero-valued activations, exploiting activation sparsity for speedup. For enabling zero-skipping, Cnvlutin incurs a mild area overhead of 4.49% over the baseline (details in [10]). Additionally, as noted in [10], Cnvlutin’s power consumption is 7% lower (on an average) than the baseline. Thus, Cnvlutin reduces latency without increasing power consumption.

Fig. 3 shows the normalized latencies of the ResNet-18 and LeNet-5 QDNNs. By exploiting the inherent activation sparsity, Cnvlutin provides a 41.6% and 32.9% latency reduction over the baseline (without zero-skipping), for ResNet-18 and LeNet-5 standard QDNNs, respectively. Further, we observe that the ResNet-18 and LeNet-5 AS-QDNNs exhibit 60.9% and 68.7% lower latency when deployed on the Cnvlutin architecture than the standard QDNNs on the baseline architecture. Importantly, the increase in the sparsity from 53% (43%) to 75% (84%) when going from ResNet-18 (LeNet-5) QDNN to AS-QDNN leads to a latency reduction of 32.9% (53.3%). Thus,sparsity enhancement directly leads to latency reduction.

III-B2 Impact of faults on Inference Accuracy

Next, we compare the fault tolerance of AS QDNNs and standard QDNNs, to gauge whether the sparsity enhancement has any effect on the accuracies of the models in the presence of faults. The dashed lines in Figures 5(a)5(f) show the impact of bit-flip, SA0 and SA1 faults on standard and AS LeNet-5 and ResNet-18 QDNNs. We observe that the AS QDNNs have lower inference accuracy in the presence of faults compared to the standard QDNNs. For bit-flip faults, the accuracy of LeNet-5 and ResNet-18 AS QDNNs is 2.40%-11.13% and 0.52%-8.00% lower (absolute difference in accuracy) than their equivalent standard QDNNs, respectively. This trend holds for SA0 and SA1 faults, albeit with lower accuracy degradation than bit-flip faults as the SA faults can potentially be masked [15]. These results signify that both AS QDNNs are less fault tolerant than their standard QDNN counterparts. This implies that the latency benefits associated with increased activation sparsity comes at the price of higher vulnerability to faults.

Refer to caption
Figure 4: (a) The weight loss landscape of the standard and AS LeNet-5 QDNNs visualized using the technique in [9]. x and y are normalized random directions. (b) and (c) show the impact of a fault (F) on the loss value in a landscape with (b) sharp and (c) flat minima. The fault causes a larger change in the loss value in the former.
Refer to caption
(a) Bit-flip faults on LeNet-5 QDNNs
Refer to caption
(b) SA0 faults on LeNet-5 QDNNs
Refer to caption
(c) SA1 faults on LeNet-5 QDNNs
Refer to caption
(d) Bit-flip faults on ResNet-18 QDNNs
Refer to caption
(e) SA0 faults on ResNet-18 QDNNs
Refer to caption
(f) SA1 faults on ResNet-18 QDNNs
Figure 5: Comparison of the impact on classification accuracy for different fault scenarios for both SAQ trained and conventionally trained standard and activation-sparse (AS) QDNNs. It can be seen that AS QDNNs have lesser fault tolerance than their standard counterparts. Also, SAQ-trained QDNNs display higher fault tolerance than their conventionally trained equivalents.

III-B3 Weight Loss Landscape visualization

To explain the reduced fault tolerance of AS QDNNs, we visualize their weight loss landscapes. Figure 4a shows the weight loss landscape of the standard and AS LeNet-5 QDNN. Intuitively, a QDNN with a sharper minima would incur a larger change in the loss function value due to weight perturbation, as shown in Figure 4b and c. We observe that the AS QDNN converges at a sharper minima, which explains its experimentally observed degraded fault tolerance. Thus, the increased activation sparsity due to the L1L1 activation regularization is accompanied by an increase in the sharpness of the minima and reduced fault tolerance. A similar trend was observed for ResNet-18.

To sum up, enhanced activation sparsity reduces the latency of sparsity-aware DNN accelerators, but degrades fault tolerance due to a sharper minima in the loss landscape. Based on this observation, we propose the utilization of sharpness-aware quantization (SAQ) based training to enhance the fault tolerance of AS QDNNs and show its effectiveness next.

IV Mitigation Strategy: SAQ

IV-A SAQ-based fault mitigation strategy

SAQ flattens the weight loss landscape of the QDNNs, which motivates us to employ it for making the QDNN less sensitive to faults. To evaluate this approach, we utilize SAQ to train both standard and AS QDNNs.

SAQ concurrently minimizes the loss value and loss sharpness by solving the following min-max optimization problem:

min𝑤maxϵ2ρ((LS(Qw(w,b))+ϵ)LS(Qw(w,b)))+LS(Qw(w,b))+λ2w22\underset{w}{\min}\underset{||\epsilon||_{2}\leq\rho}{\max}((L_{S}(Q_{w}(w,b))+\epsilon)-L_{S}(Q_{w}(w,b)))\\ +L_{S}(Q_{w}(w,b))+\frac{\lambda}{2}||w||^{2}_{2} (2)

The first term in the equation defines the sharpness metric, which is the maximum change in the loss value for some weight perturbation ϵ\epsilon. The second term is the loss function itself and the third term is the standard L2L2 regularization term. Also, the perturbation ϵ\epsilon is an adversarial one and is chosen such that the maximum sharpness is minimized. ϵ\epsilon is given by the following equation and can have a maximum value of ρ\rho.

ϵϵ^=ρQw(w,b)LS(Qw(w,b))Qw(w,b)LS(Qw(w,b))2\epsilon\simeq\hat{\epsilon}=\rho\frac{\nabla_{Q_{w}(w,b)}L_{S}(Q_{w}(w,b))}{||\nabla_{Q_{w}(w,b)}L_{S}(Q_{w}(w,b))||_{2}} (3)

Here, Qw(w,b)LS(Qw(w,b))\nabla_{Q_{w}(w,b)}L_{S}(Q_{w}(w,b)) is the gradient of the loss function with respect to the quantized weights. ϵ\epsilon is estimated using a forward and backward pass through the QDNN. Thus, compared to conventional training, SAQ training requires double the number of forward/backward passes per epoch. For a fair comparison of SAQ-based QDNNs with conventionally trained ones, we train SAQ models for half as many epochs. We select the ρ\rho hyperparameter to maximize the fault-free accuracy (fault rate of 0%) for the QDNNs. Using this framework, we analyse the effectiveness of SAQ in increasing the fault tolerance for both standard QDNNs and AS QDNNs.

IV-B Results

First, we compare standard QDNNs trained conventionally and with SAQ. Figure 5 shows that SAQ-trained standard QDNNs exhibit larger fault tolerance than their conventionally trained counterparts. For example, for a bit-flip fault rate of 3%, SAQ-training increases the inference accuracy of LeNet-5 and ResNet-18 by 6.70% and 12.49%, respectively.

Next, we compare the accuracies of SAQ-trained and conventionally trained AS QDNNs. The L1L1 regularization constant is kept the same for both SAQ-trained and conventionally trained AS QDNNs, ensuring similar activation sparsity. The trends observed for AS QDNNs are similar to those for standard QDNNs. For a bit-flip fault rate of 3%, SAQ training enhances the inference accuracy of LeNet-5 and ResNet-18 AS QDNNs by 9.34% and 13.96%.

An interesting observation here is that SAQ-trained AS QDNNs exhibit a higher accuracy in the presence of faults than standard QDNNs without SAQ. For example, our experiments show that the SAQ-trained LeNet 5 AS QDNN has a 8.37% higher accuracy than the conventionally trained LeNet 5 standard QDNN at fault rate = 5%. Thus, SAQ-trained AS QDNNs have both the benefits of superior fault tolerance and lower latency (compared to conventionally trained standard QDNNs), making them highly suitable for edge applications. Thus, training with SAQ mitigates the adverse impact that activation sparsity augmentation has on the fault tolerance.

V Conclusion

In this work, we show that the increase in activation sparsity (and the associated speedup) comes at the price of degraded inference accuracy in the presence of faults, with activation sparse (AS) QDNNs showing up to 11.13% lower accuracy than standard QDNNs. To explain this reduced fault tolerance, we visualize the weight loss landscapes for standard and AS QDNNs and establish that AS QDNNs exhibit a sharper minima leading to lower fault tolerance. Based on this observation, we employ sharpness-aware quantization (SAQ) based training, which flattens the loss landscape and enhances fault tolerance. Our results show that AS and standard QDNNs trained with SAQ have up to 19.50% and 15.82% higher accuracy than their conventionally trained counterparts. We also observe that SAQ-trained AS QDNNs have higher inference accuracy than conventionally trained standard QDNNs, enabling QDNNs which are not only activation-sparse (and thus, low-latency) but also fault-tolerant.

References

  • [1] W. Li et al., “A precision-scalable deep neural network accelerator with activation sparsity exploitation,” IEEE Trans. Comp.-Aid. Des. Integ. Circ. Sys., 2024.
  • [2] B. Jacob et al., “Quantization and training of neural networks for efficient integer-arithmetic-only inference,” in 2018 IEEE/CVF Conf. Comp. Vis. Pat. Rec., 2018, pp. 2704–2713.
  • [3] S. Vadera and S. Ameen, “Methods for pruning deep neural networks,” IEEE Access, vol. 10, pp. 63 280–63 300, 2022.
  • [4] M. Kurtz et al., “Inducing and exploiting activation sparsity for fast inference on deep neural networks,” in Proc. 37th Int. Conf. Mach. Learn., 2020, pp. 5533–5543.
  • [5] G. Georgiadis, “Accelerating convolutional neural networks via activation map compression,” in 2019 IEEE/CVF Conf. Comp. Vis. Pat. Rec. (CVPR), 2019, pp. 7078–7088.
  • [6] A. Bhattacharjee et al., “Examining and mitigating the impact of crossbar non-idealities for accurate implementation of sparse deep neural networks,” in Proc. 2022 Conf. & Exhib. Des., Auto. & Test Eur., 2022, p. 1119–1122.
  • [7] G. Yuan et al., “Improving dnn fault tolerance using weight pruning and differential crossbar mapping for reram-based edge ai,” in 2021 22nd Int. Symp. Qual. Elec. Des. (ISQED), 2021, pp. 135–141.
  • [8] J. Liu et al., “Sharpness-aware quantization for deep neural networks,” 2023.
  • [9] H. Li et al., “Visualizing the loss landscape of neural nets,” in Adv. Neur. Inf. Proces. Sys., vol. 31, 2018.
  • [10] J. Albericio et al., “Cnvlutin: Ineffectual-neuron-free deep neural network computing,” in 2016 ACM/IEEE 43rd Ann. Int. Symp. Comp. Arch. (ISCA), 2016, pp. 1–13.
  • [11] S. Mittal, “A survey on modeling and improving reliability of dnn algorithms and accelerators,” Jour. Sys. Arch., vol. 104, p. 101689, 2020.
  • [12] F. Libano et al., “Selective hardening for neural networks in fpgas,” IEEE Trans. Nuc. Sci., vol. 66, no. 1, pp. 216–222, 2019.
  • [13] A. Malhotra et al., “Tfix: Exploiting the natural redundancy of ternary neural networks for fault tolerant in-memory vector matrix multiplication,” in 2023 60th ACM/IEEE Des. Aut. Conf. (DAC), 2023, pp. 1–6.
  • [14] T. Liu et al., “A fault-tolerant neural network architecture,” in 2019 56th ACM/IEEE Des. Aut. Conf. (DAC), 2019, pp. 1–6.
  • [15] A. Malhotra et al., “Bnn-flip: Enhancing the fault tolerance and security of compute-in-memory enabled binary neural network accelerators,” in 2024 29th Asia S. Pac. Des. Aut. Conf. (ASP-DAC), 2024, pp. 146–152.
  • [16] B. Zhang et al., “Handling stuck-at-fault defects using matrix transformation for robust inference of dnns,” IEEE Trans. Comp.-Aid. Des. Integ. Circ. Sys., vol. 39, no. 10, pp. 2448–2460, 2020.
  • [17] U. Zahid et al., “Fat: Training neural networks for reliable inference under hardware faults,” in 2020 IEEE Int. Test Conf, 2020, pp. 1–10.
  • [18] Y. Jiang et al., “Fantastic generalization measures and where to find them,” in Int. Conf. Learn. Repres., 2020.
  • [19] P. Foret et al., “Sharpness-aware minimization for efficiently improving generalization,” in Int. Conf. Learn. Repres., 2021.