Next Article in Journal
Reliability and Detectability of Emergency Management Systems in Smart Cities under Common Cause Failures
Previous Article in Journal
A Lightweight Remote Sensing Small Target Image Detection Algorithm Based on Improved YOLOv8
Previous Article in Special Issue
Prototype Learning for Medical Time Series Classification via Human–Machine Collaboration
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improving Adversarial Robustness of ECG Classification Based on Lipschitz Constraints and Channel Activation Suppression

1
School of Electronic and Information Engineering (SEIE), Zhuhai College of Science and Technology, Zhuhai 519041, China
2
College of Communication Engineering, Jilin University, Changchun 130012, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(9), 2954; https://doi.org/10.3390/s24092954
Submission received: 8 April 2024 / Revised: 28 April 2024 / Accepted: 1 May 2024 / Published: 6 May 2024
(This article belongs to the Special Issue Sensors Technology and Application in ECG Signal Processing)

Abstract

:
Deep neural networks (DNNs) are increasingly important in the medical diagnosis of electrocardiogram (ECG) signals. However, research has shown that DNNs are highly vulnerable to adversarial examples, which can be created by carefully crafted perturbations. This vulnerability can lead to potential medical accidents. This poses new challenges for the application of DNNs in the medical diagnosis of ECG signals. This paper proposes a novel network Channel Activation Suppression with Lipschitz Constraints Net (CASLCNet), which employs the Channel-wise Activation Suppressing (CAS) strategy to dynamically adjust the contribution of different channels to the class prediction and uses the 1-Lipschitz’s distance network as a robust classifier to reduce the impact of adversarial perturbations on the model itself in order to increase the adversarial robustness of the model. The experimental results demonstrate that CASLCNet achieves A C C r o b u s t scores of 91.03% and 83.01% when subjected to PGD attacks on the MIT-BIH and CPSC2018 datasets, respectively, which proves that the proposed method in this paper enhances the model’s adversarial robustness while maintaining a high accuracy rate.

1. Introduction

Arrhythmias are a significant group of cardiovascular diseases that can cause sudden cardiac death and pose a major threat to human health [1]. The electrocardiogram (ECG) is a diagnostic tool used to non-invasively record the heart’s electrical signals. A physician typically studies and analyzes the ECG to identify the type of disease in which it was collected [2]. However, the diagnosis requires subjective judgement by doctors with extensive clinical experience, which not only consumes a large amount of healthcare resources but also does not guarantee reliability. Therefore, researchers have begun to explore the application of efficient and accurate deep neural networks (DNNs) in the field of ECG disease diagnosis and have achieved remarkable results. Oh et al. [3] proposed a novel automated system, which achieved a 98.10% accuracy in five MIT-BIH categories. Wang et al. [4] proposed an arrhythmia classification algorithm based on the multi-head self-attention mechanism (ACA-MA) and achieved a 99.4% accuracy in five categories of the MIT-BIH dataset. Kim et al. [5] adopted a residual network with a squeeze-and-excitation (SE) block and a bidirectional long short-term memory (BIL-LSTM) for arrhythmia classification and used the synthetic minority oversampling technique (SMOTE) to solve the data imbalance, and gained a 99.20%, 99.35%, and 97.05% accuracy in MITDB, AFDB, and Cinc DB, respectively. Kumar et al. [6] built a method to extract ECG features using continuous wavelet changes and used a model with SENet and lightweight context transform (LCT) for arrhythmia classification. Zeng et al. [7] proposed Fuzz-ClustNet, which use fuzzy clustering and deep learning for ECG signals detecting arrhythmia, and achieved a 98.66% and 95.79 accuracy in the MIT-BIH and PTB dataset. Recent studies have highlighted the severe threat posed by adversarial attacks to the security of DNNs, substantiated across various domains [8,9]. Adversarial examples introduce minor perturbations to the natural ECG signals, which can cause DNNs to produce erroneous results in medical diagnoses. This can potentially lead to significant medical accidents. The authors in [10] demonstrated the deceptive nature of the electrocardiogram and introduced a novel ‘cross-subject attack’. This method uses captured victim electrocardiogram short templates to map an attacker’s electrocardiogram onto the victim’s, enabling cross-device attacks with an exceptional efficacy. Chen et al. [11] conducted a study on adversarial attacks on DNN-based ECG classification systems. They proposed two attack methods based on ECG signal characteristics and introduced a smoothness metric to quantify human-perceived distances in ECG signals. In [12], generative adversarial networks were used to create fake ECG signals using victim ECG templates. Han et al. [13] proposed the Smooth Adversarial Perturbation (SAP) method, a technique specifically designed to attack ECG signal classifiers. This method applies Gaussian kernel convolution to smooth adversarial perturbations, reducing the occurrence of physiologically implausible square-wave artefacts that may arise.
The adversarial attack algorithms applied in the field of image recognition can similarly be utilized to target ECG signals. The Fast Gradient Sign Method (FGSM), proposed by Goodfellow et al. [14], generates adversarial examples based on the model gradient and single-step optimization, representing a classic adversarial attack technique. Building upon the FGSM, Madry et al. [15] proposed the Projected Gradient Descent (PGD) adversarial attack algorithm. This method involves multiple iterations; a random perturbation not exceeding the specified perturbation range is superimposed on the natural example, and this is used as the initial adversarial examples for multiple iterations. The definition of PGD is provided in Equation (1):
x a d v = C l i p x , ϵ ( x a d v + δ     s i g n x J x a d v , y )
where x is the natural example and y is the natural example label, x a d v denotes the initial adversarial examples and x a d v is the new adversarial examples,   ε is the maximum adversarial perturbation, δ is the step size of each iteration, J ( , ) calculates the predicted loss of the neural network, and C l i p x , ε (   ) limits the size of the perturbation to the inside of the circle centred on the data x , with ε as the threshold. PGD can generate the strongest adversarial examples within the approximate sample space and stands as one of the widely used adversarial attack methods. Carlini et al. [16] proposed the C&W adversarial attack algorithm, treating adversarial examples as optimizable variables. They designed a loss function to transform the generation process of adversarial examples into a solvable optimization problem. Currently, C&W is regarded as one of the most effective white-box attack algorithms based on gradient optimization.
In order to safeguard DNNs from malicious attacks using adversarial examples on ECG signals, researchers have conducted in-depth investigations. Wiedeman et al. [17] introduced a novel ensemble method based on feature decorrelation and Fourier partitioning to enhance network features and reduce the impact of adversarial attacks. This approach aims to fortify the network against adversarial perturbations. Jeong et al. [18] proposed Defensive Adversarial Training, which involves training the model using diversified noise data to enhance the robustness of the recognition algorithm. The results demonstrate the significant effectiveness of this method in resisting noise injection and random noise compared to traditional noise removal solutions. To enhance the robustness of ECG signal classification models against adversarial noise, Ma et al. [19] introduced a regularization method based on the Noise-to-Signal Ratio (NSR). The approach aims to improve the robustness of DNNs against adversarial perturbations. Shao et al. [20] proposed a defense method based on adversarial distillation training, demonstrating its efficacy in enhancing the generalization performance of DNNs against adversarial attacks in ECG classification. The above literature does not explore the effect of the model structure on the robustness of ECG signal classification. However, recent research highlights the critical roles played by the feature extraction module and classifier in adversarial robustness [21,22,23]. In response, this research focuses on elucidating the influence of the model structure on the adversarial robustness of ECG signals. Our efforts are directed towards enhancing the model architecture, particularly the channel activation in the feature extraction stage and the design of the classifier, to improve the model’s robustness against adversarial examples in ECG signals. During the feature extraction phase, adversarial perturbations accumulate distortions in the channel activation magnitude, leading to a signal enhancement effect that renders the model prone to misclassification under adversarial attacks. Moreover, for natural examples of the same category, robust channels in the model generate more universally applicable patterns, whereas adversarial examples frequently activate non-robust channels, resulting in incorrect model outputs and diminishing network robustness [24]. In the feature extraction stage, the primary distinctions in features between adversarial and natural examples originate from variations in the channel activation magnitude and frequencies induced by adversarial perturbations, thus influencing the model’s performance in adversarial robustness. During the classification stage, adversarial attacks on the feature extraction phase induce variations in feature vectors, consequently leading to misclassifications by the classifier. Lipschitz continuity imposes a constant constraint on the range of variations between the input adversarial perturbation and the output of the model, with the minimum non-negative constant satisfying this property referred to as the Lipschitz constant for the classifier [25]. By designing a classifier with a Lipschitz constant constraint, the classification accuracy of adversarial examples can be effectively improved, particularly those with significant differences from the feature vectors of natural examples. This constraint plays a positive role in reinforcing the model’s stability and adversarial robustness.
The main contributions of this study are summarized as follows:
  • We proposed a novel robust model Channel Activation Suppression with Lipschitz Constraints Net (CASLCNet). In the feature extraction stage, CASLCNet employs the Channel-wise Activation Suppressing (CAS) strategy with an auxiliary classifier for the adaptive learning of the channel importance. This strategy dynamically adjusts channels to suppress non-robust channels. In the classification stage, CASLCNet utilizes a distance network with the Lipschitz continuity as the classifier, effectively resisting small perturbations generated by adversarial attacks.
  • We employed Misclassification Aware Adversarial Training (MART), which can further improve the adversarial robustness of CASLCNet for ECG classification.
  • We validated the model adversarial robustness using multiple adversarial attack methods in the MIT-BIH dataset and the CPSC2018 dataset and compare it with state-of-the-art methods. The experimental results show that the method in this paper can effectively defend against malicious attacks on the model by multiple adversarial attack methods while maintaining a high accuracy, and outperforms the state-of-the-art methods in a variety of metrics.

2. Materials

2.1. Datasets

This study validates the performance of the CASLCNet network using the processed MIT-BIH dataset [26,27] and the CPSC2018 dataset [28]. The specific details of the dataset are presented in Table 1.
  • The MIT-BIH dataset, sampled at a frequency of 125 Hz, is categorized into five classes based on AAMI standards, resulting in a total of 109,446 samples.
  • The CPSC2018 dataset, sampled at a frequency of 500 Hz with a duration ranging from 6 to 60 s, comprises nine sample classes. To streamline our experimentation, 477 electrocardiogram samples with multiple labels are excluded, leaving 6400 samples of 12-lead electrocardiograms for further analysis.

2.2. Validation Metrics

This study employs two fundamental evaluation metrics: accuracy and F1 score. Additionally, A C C r o b u s t and F 1 r o b u s t [19], as two evaluation metrics, are utilized to measure the overall robustness of the network within a specific attack range. The definition of A C C r o b u s t is provided in Equation (2):
A C C r o b u s t = A C C c l e a n × A U C ϵ ϵ m a x
Additionally, A C C c l e a n represents the normalized classification accuracy, and A U C | | ϵ | | ϵ m a x is the normalized area under the curve in the presence of varying noise levels. It calculates the average classification accuracy within a certain range of noise levels | | ϵ | | ϵ m a x . Therefore, A C C r o b u s t is able to provide a comprehensive performance indicator that measures both a clean sample accuracy and overall robustness. Similarly, the definition of another metric F 1 r o b u s t is given by Equation (3):
F 1 r o b u s t = F 1 c l e a n × A U C ϵ ϵ m a x
Due to the potential significant impact of small adversarial perturbations on the model output, a maximum perturbation ϵ m a x is set to consider only a specific range of perturbation | | ϵ | | ϵ m a x . In this experiment, for the MIT-BIH dataset with ϵ m a x   of 0.3, the ranges of perturbation are set to 0.01 , 0.03 , 0.05 , 0.1 , 0.2 , 0.3 and the number of iterations is set to 100. In the CPSC2018 dataset, ϵ m a x   is 0.1 and the ranges of perturbation are set to 0.001 , 0.003 , 0.005 , 0.007 , 0.01 , 0.03 , 0.05 , 0.1 and the number of iterations is set to 100. In the SAP attack algorithm, gaussian kernel with size s is set to   5 , 7 , 11 , 15 , 19 , and standard deviation σ is set to 1.0 , 3.0 , 5.0 , 7.0 , 10.0 .

3. Methods

In this section, the adversarial robustness model for arrhythmia classification using CASLCNet is described in detail, and the overall flowchart is shown in Figure 1. Firstly, the ECG signals in the dataset are preprocessed and initialized to generate adversarial examples; the natural examples and the adversarial examples are fed into CASLCNet for feature extraction to obtain the feature vectors and the auxiliary classifier prediction probabilities, respectively, and the feature vectors are fed into the distance network as the classifier to obtain the prediction probabilities. The above data are valued for adversarial training, i.e., maximizing the Cross-Entropy loss function to find the worst-case samples, minimizing the adversarial loss function to train a model that is robust to the adversarial examples, and constantly updating the model network parameters while updating the adversarial examples for training, in the hope of obtaining a model with better adversarial robustness.

3.1. Data Preprocessing

Preprocessing of ECG data is required before model training. The MIT-BIH dataset needs to perform up-sampling operations on the training and test sets to eliminate the effect of class imbalance on model training. The CPSC2018 dataset needs to populate the ECG data less than 60 S with 0 to 60 S at both ends, and the ECG data more than 60 S with the first 60 S intercepted. Since leads V3, V4, V5, and V6 can be obtained from other leads, these four leads are removed and the remaining eight leads are saved. Using feature normalization, the lead values are deflated to between −1 and 1 according to the maximum value of each lead, again using up-sampling to eliminate the effect of category imbalance.

3.2. Feature Extraction and Classification (CASLCNet)

In this section, CASLCNet is described in detail, and the general framework is shown in Figure 2. The network consists of residual modules [29], residual modules with Channel-wise Activation Suppression strategy [30], and distance network [31]. In the feature extraction stage, the residual module and the residual module with Channel-wise Activation Suppression strategy are used to extract the deep features of the ECG signal while dynamically adjusting the channel importance. Meanwhile, using the distance network as the classifier can constrain the Lipschitz constant of the classifier, which can effectively resist the influence of adversarial perturbations on the model output. The details of the model will be introduced one by one in the following. Since the data in the CPSC2018 dataset are variable-length data, zero padding is used in the preprocessing stage to fix the length of the ECG signal to make it easier to feed into the model for computation. To eliminate the effect of zero padding in the lead data on the model, average pooling is performed using a fixed-length input mask in the dimensionality reduction stage in CASLCNet and multiplied with the extracted feature vectors to obtain a valid output vector. After the output masking operation, channel averaging weighted by the mask is performed to reduce the dimensionality of the output vectors from variable-length to a fixed length, which is then fed into the distance network for classification. The specific flowchart is shown in Figure 3. Dimension reduction is carried out using global average pooling and dimension transformation in the MIT-BIH dataset.

3.2.1. Residual Module

The expressive capability of DNNs gradually increases with the addition of layers, allowing for higher accuracy and stronger performance through deeper network architectures. However, the augmentation of depth in neural networks poses challenges such as overfitting, gradient explosion, gradient vanishing, and a decline in information propagation capability, leading to a degradation in model performance. In addressing these challenges, He et al. [29] introduced the concept of residual learning into deep neural networks and proposed residual network ResNet.
The residual module used in this paper is modified from the original residual module, consisting of convolutional layers, group normalization layers, activation functions, and a skip connection, is designed to learn deeper-level features. This module adds the input of the module to the output obtained after convolutional operations. This mechanism facilitates the transfer of gradient information between layers, overcoming issues like gradient vanishing. Consequently, the model’s overall performance is enhanced.

3.2.2. Channel-Wise Activation Suppression Strategy

During the model training phase, the Channel-wise Activation Suppression strategy dynamically learns the importance of channels. By suppressing non-robust channels, it effectively reduces the activation magnitude of feature vector channels and the frequency of network channel activations. Let the k-th CAS module receive the feature vector activated by the residual module through the ReLU function f k R W × C , where C represents the number of channels in the feature vector, and W is the width of the feature vector. Initially, the feature vector f k is subjected to global average pooling to obtain the channel activation vector f ^ k . Subsequently, this vector f ^ k is fed into a fully connected layer for category classification. Assuming there are M classes in the dataset, the parameters of the fully connected layer for the auxiliary classifier can be expressed as follows: H k = [ H 1 k , H 2 k , . . , H C k ] R C × M . The fully connected layer identifies the importance of each channel for a given class, and reweights the natural feature vector accordingly, followed by forward propagation to the next layer.
During the model training phase, the heartbeat data label y is used as an indicator for determining the importance of classes, denoted as H y k R c . As obtaining data label information is not possible during the testing phase, the predicted class by the auxiliary classifier is used as channel weights, denoted as H y ^ k k R C . For adversarial examples x a d v , the output probability of the auxiliary classifier is p ^ k = s o f t m a x ( f ~ k H k ) R M . Let K represent the total number of CAS modules in the network, and the loss function used for network training is L . The overall loss function for CAS modules during network training is defined in Equation (4):
L C A S x , x a d v , y ; θ , H = α K k = 1 K   L p ^ k x , x a d v , θ , H , y
Among other things, θ denotes the model parameters, and α is the hyperparameter that balances the training of CAS modules.

3.2.3. Distance Network

The distance network is a multilayer perceptron network composed of distance neurons as fundamental units. distance neurons take the feature vector x f as input, with an added bias term b . The parametric distance can be computed by the norm distance between x f and the parameter w , as defined by Equation (5):
u x f , w , b = x f w + b
Based on the definition of distance neurons, a fully connected distance network can be constructed. The distance network module g takes x f ( 0 ) = x f as input, and its L -th layer, x f ( l ) , is defined as Equation (6):
x f i l = u x f l 1 , w l , j , b i l = x f l 1 w l , i + b i l , l L , i n l
Here, n l represents the number of neurons in the l -th layer. For a classification problem with M classes, n L = M . The l distance network module takes g ( x f ) = x f ( L ) as output probabilities and predicts the class a r g m a x i [ M ]   [ g ( x f ) ] i . Due to the 1-Lipschitz mapping property of the distance layer concerning norms, any distance network is inherently 1-Lipschitz through composition.

3.3. Misclassification-Aware Adversarial Training

In this study, we employed Misclassification Aware Adversarial Training [32] to train the network. For an M-class classification problem, we are given a dataset { ( x i , y i ) } i = 1 , , n , where the natural example x i R d and y i { 1 , , M } represent the class. For a deep neural model h θ with network parameters θ , an adversarial sample x a d v is generated based on the natural example x i . During adversarial training, both the natural examples and adversarial examples are fed into the h θ to obtain model prediction probability values. Samples with model prediction errors are classified into three categories: natural example prediction error h θ ( x i ) y i , adversarial sample prediction error h θ ( x a d v ) y i , and inconsistency in predictions between the natural examples and adversarial examples h θ ( x i ) h θ ( x a d v ) . MART addresses these three types of misclassifications during adversarial training, and its loss function is defined by Equation (7):
L MART x , x a d v , y ; θ = λ K L p x i , θ p x a d v , θ 1 p y i x i , θ     + 1 n i = 1 n   B C E p x a d v , θ , y i
Here, boosted cross-entropy (BCE) loss is used for misclassifications where h θ ( x a d v ) y i , allowing the model to obtain a stronger decision boundary. For misclassifications where h θ ( x i ) h θ ( x a d v ) , KL divergence is employed to minimize the distribution difference between the two, thereby better fitting the outputs of the natural examples and adversarial examples. Regarding misclassifications where h θ ( x i ) y i , the soft decision ( 1 p y i ( x i , θ ) ) dynamically adjusts the loss function size for improved robust network training. In the training of the model in this study, MART is also applied to the auxiliary classifier. Therefore, the overall loss function during network training is given by Equation (8):
L = L MART x , x a d v , y ; θ + α R r = 1 R   L C A S MART x , x a d v , y ; θ , H
The direct use of MART for the CPSC2018 dataset leads to difficulties in model convergence. Therefore, it is necessary to add the MSE loss function to assist the model convergence during the training process. The overall loss function formula is given by Equation (9):
L = L MART x , x a d v , y ; θ + α R r = 1 R   L C A S MART x , x a d v , y ; θ , H + 1 n i = 1 n   ( y i h θ x i ) 2

4. Result and Discussion

4.1. Experimental Setup

The research experiments are conducted on a server equipped with an Intel(R) Xeon(R) Gold 5218 CPU (2.30 GHz) and NVIDIA A100-SXM4 GPU (40 GB memory). The operating system used is Centos 8, with Python version 3.8.3, PyTorch version 1.13.1, and CUDA version 11.6. For the MIT-BIH dataset, the batch size is set to 512, the number of training rounds is 100, the Adamax optimizer is used for training, the initial learning rate is set to 1 × 10−3, and the ReduceLROnPlateau learning rate scheduler is used to dynamically adjust the learning rate. For CASLCNet training, the PGD attack algorithm is used to generate adversarial examples for adversarial training, the number of attacks is set to 10, the attack range is 0.1, and the attack step size is one-tenth of the attack range. For the CPSC2018 dataset, the batch size during training is set to 64, the initial learning rate is set to 1 × 10−4, and the model is also trained using MART with the number of attacks set to 10, the attack range to 0.01, and the rest of the settings the same as for the MIT-BIH dataset. Table 2 shows the detailed structure of the network using CASLCNet for the MIT-BIH and CPSC2018 dataset.

4.2. Channel-Wise Activation Suppression Effect

As examples, we chose the N class and Normal class from the MIT-BIH and CPSC2018 dataset. Our observation focused on the channel activation frequency and magnitude at the final layer of the model’s feature extraction. For each channel, if the activation value surpassed a threshold (20% of the maximum activation value across all 512 channels in MIT-BIH, and 70% of the maximum activation value in CPSC2018), the channel is identified as an activated channel. Subsequently, we calculated the activation frequency on each channel for both the natural examples and adversarial examples, sorting them in descending order of the natural example’s activation frequency. In the experiments, CASLCNet is trained using MART and ResNet18 is trained using the cross-entropy loss function as the contrast model. Figure 4 and Figure 5 illustrates the channel-wise activation frequency and magnitude of ResNet18 and CASLCNet on the test sets of both datasets. From the subfigures a, it is evident that the channel-wise activation magnitude of the adversarial examples is significantly higher than that of the natural examples. This indicates that adversarial perturbations progressively accumulate from the model’s input layer to the output layer. By looking at the subfigures c, we notice that adversarial examples activate the model channels more uniformly, frequently activating non-robust channels seldom activated by the natural examples. This has a severe impact on the model’s robustness. Subfigures b depict the activation magnitude of CASLCNet when faced with adversarial examples. It is apparent that our proposed method effectively suppresses the activation magnitude of adversarial examples, reducing the magnitude gap between adversarial and natural examples. Subfigures d represent the channel-wise activation frequency of adversarial examples. Our proposed method effectively suppresses the channel activation frequency, aligning the activation frequencies of natural examples and adversarial examples and reducing the activation on non-robust channels by adversarial examples. Consequently, this mitigates the impact of adversarial sample attacks on the network, enhancing overall robustness.

4.3. Hyperparameter Selection Experiment

In the training of the CASLCNet network, adjustments to the α parameter of the CAS loss are made to achieve optimal training outcomes. To assess the sensitivity of the CAS strategy under different α values, MART is conducted on the MIT-BIH and CPSC2018 datasets for α values of 0 , 1 , 2 , 3 , 4 , where α = 0 represents standard adversarial training. Table 3 presents the corresponding A C C r o b u s t and   F 1 r o b u s t scores for each α value. The results indicate that the model achieves optimal performance across metrics when α is set to 2, striking a balance between the accuracy rate and robustness. Figure 6 shows the loss function curves as well as the accuracy curves of CASLCNet in the MIT-BIH dataset and the CPSC2018 dataset when the hyperparameter α is 2.

4.4. Adversarial Robustness Verification

To assess the effectiveness of CASLCNet in defending against various malicious attacks, we conducted validation using different adversarial attack methods on the test sets of the MIT-BIH and CPSC2018 datasets. The adversarial attack methods employed included white noise, FGSM, C&W, PGD, and SAP. White noise and FGSM attacks utilized a single iteration, while MI-FGSM and C&W used 100 iterations. White noise and FGSM and MI-FGSM perturbation ranges are set to 0.1 and 0.01 in the MIT-BIH and CPSC2018 datasets, respectively, with an MI-FGSM step range of 0.01 and 0.001. When using PGD and SAP adversarial attacks, the settings are as shown in 2.2. Table 4 provides the detailed accuracy and F1 scores under the white noise, FGSM, MI-FGSM, and C&W attack methods. Table 5 and Table 6 show the accuracy and F1 scores of CASLCNet when the MIT-BIH dataset and the CPSC2018 dataset are attacked by different PGD and SAP adversarial attack, respectively. Notably, the model’s accuracy and F1 scores showed minimal degradation when faced with white noise, FGSM, MI-FGSM, C&W, and SAP attacks. Even under PGD adversarial attacks, the model maintained a high level of accuracy, demonstrating the model’s ability to effectively withstand various adversarial attacks while preserving a high accuracy.

4.5. Ablation Experiment

To assess the effectiveness of each module in enhancing the adversarial robustness of the CASLCNet network, we conducted ablation experiments on both the MIT-BIH and CPSC2018 datasets. Method 1 employed ResNet18 as the baseline model, while Method 2 replaced the last feature extraction layers of Method 1 with residual modules incorporating the channel-wise activation suppression strategy. Method 3 replaced the fully connected layer of Method 1 with an distance network serving as the classifier. Method 4 represents our proposed CASLCNet. All methods utilized MART. Table 7 shows the A C C r o b u s t and F 1 r o b u s t scores under the test set of MIT-BIH and CPSC2018 datasets, where × means that the method model does not contain the module, and √ means that the model contains the module, and it can be observed that Method 4 achieves the best values in all the metrics; it shows that both modules added in this paper are effective in improving the model adversarial robustness.

4.6. Contrast Experiment

In this study, CASLCNet is trained using various methodologies, including standard adversarial training [15], TRADES adversarial training [33], and MART. The experimental results are presented in Table 8. Table 8 demonstrates that Misclassification-Aware Adversarial Training consistently achieves optimal values across metrics. To verify the effectiveness of the CASLCNet model, three classical networks, VGG19 [34], ResNet18 and DenseNet [35], are used in this paper and the proposed CASLCNet is trained with different loss functions, respectively, and the experimental results are shown in Table 9 and Table 10, respectively. Table 9 shows the detailed results of the A C C r o b u s t and F 1 r o b u s t scores of each method in the MIT-BIH dataset, and Table 10 shows the detailed results of the A C C r o b u s t and F 1 r o b u s t scores of each method in the CPSC2018 dataset. Table 10 shows that the A C C r o b u s t and F 1 r o b u s t scores of CASLCNet are higher when the model is attacked by adversarial examples when trained with the cross-entropy loss function, compared to the A C C r o b u s t and F 1 r o b u s t scores of VGG19, ResNet18, and DenseNet trained with MART, which effectively shows that the CASLCNet network proposed in this paper has strong adversarial robustness without adversarial training. The A C C r o b u s t and F 1 r o b u s t scores of CASLCNet under PGD adversarial attack can be improved by more than 40% compared to other methods when CASLCNet is trained using MART in the CPSC2018 dataset. Table 9 and Table 10 show that the proposed method can achieve the best performance in each index, and can effectively improve the A C C r o b u s t and F 1 r o b u s t scores compared with the other networks, which indicates that CASLCNet can achieve the advantage in the accuracy of the natural examples, as well as the robustness. This observation substantiates the effectiveness of CASLCNet in significantly enhancing the model’s adversarial robustness.

4.7. Comparison with Existing Literature

Recent studies have indicated that SNR regularization enhances network robustness by suppressing the Signal-to-Noise Ratio (SNR) of adversarial noise signals, while Jacobian regularization mitigates the impact of adversarial noise perturbations by penalizing large gradients relative to the output. These regularization methods represent advanced approaches for defending against adversarial attacks on electrocardiographic signals. Figure 7 and Figure 8 show the histograms of the accuracy and F1 scores of the proposed method with the two methods, Jacob, as well as SNR, under different PGD and SAP adversarial attacks. Table 11 shows the detailed data of the comparison between the proposed method and the existing literature, from which it can be seen that the proposed method in this paper achieves the optimal results in terms of A C C r o b u s t and F 1 r o b u s t scores. In the CPSC2018 dataset, when attacked by PGD, the A C C r o b u s t and F 1 r o b u s t scores of this paper’s method are more than 30% higher than other methods. In the MIT-BIH dataset, this paper’s method is also reaching the best index. The above experiments fully prove that the proposed method in this paper achieves a better balance between identifying natural examples and adversarial examples; not only can it maintain a high accuracy, but it can also effectively resist the attack of adversarial examples, which effectively indicates that the proposed method in this paper is better than the existing methods in the literature and reaches the advanced level.
This study aims to investigate the effectiveness of the proposed CASLCNet in defending against adversarial examples when applied to medical diagnostics using electrocardiographic signals. We conducted experiments focusing on Channel-wise Activation Suppression, hyperparameter selection, robustness validation, and comparisons with existing literature.
In our observations, CASLCNet demonstrated significant advantages when confronted with various adversarial attack methods on the test sets of the MIT-BIH and CPSC2018 datasets. Under white noise, FGSM, MI-FGSM, C&W, PGD, and SAP attacks, CASLCNet maintained a high accuracy and F1 scores, showcasing its robust resistance to diverse adversarial attack methods. Furthermore, we conducted an in-depth investigation into the efficacy of the Channel-wise Activation Suppression strategy within CASLCNet. By scrutinizing the channel activation frequencies and magnitude of CASLCNet in the MIT-BIH and CPSC2018 datasets, we observed a significant reduction in the activation magnitude of adversarial examples. This reduction resulted in a diminished Magnitude gap between adversarial and natural examples, indicating a pivotal role played by the Channel-wise Activation Suppression strategy in effectively enhancing the model’s robustness. In terms of model training, we used an adversarial training approach that emphasizes misclassification. This approach involves designing the loss function to address misclassifications of both natural examples and adversarial examples. The results demonstrated that CASLCNet consistently achieved a favorable performance on the MIT-BIH and CPSC2018 datasets. These experiments serve as empirical evidence of the effectiveness of proposed method in enhancing the model’s robustness. Comparisons with the existing literature demonstrated that CASLCNet consistently achieved the best results in terms of A C C r o b u s t and F 1 r o b u s t scores, showcasing its significant advantage in adversarial attacks. This establishes CASLCNet as an advanced technology in the field of robustness research.
The proposed method in this paper achieves a significant robustness improvement in the context of arrhythmia classification, which can be applied to the automatic diagnosis of arrhythmia models, and can effectively prevent attackers from causing misdiagnosis leading to medical accidents by formulating specific adversarial examples to deceive the model, which is of great significance for improving the reliability and safety of ECG signal processing systems in practical medical applications. However, there are still some problems that need to be improved in the method of this paper. For multi-lead ECG signals, this paper’s method does not consider the influence of each lead’s signals on the robustness of the model, and there is the problem of the long computation time and large computation volume. In future work, we will investigate other applications in the field of ECG signal classification, such as identity recognition, and consider methods such as introducing a lead attention mechanism to further investigate the effect of each lead on the adversarial robustness of the model, as well as lightening the modules in CASLCNet by optimizing the adversarial training algorithms to reduce the time and computational volume.

5. Conclusions

In order to enhance the robustness of the electrocardiographic signal classification model, this paper proposes a novel robust network, CASLCNet. By leveraging a Channel-wise Activation Suppression strategy, CASLCNet dynamically adjusts the importance of channels, reducing the activation frequency of non-robust channels. Simultaneously, the introduction of an distance network composed of distance neurons serve as the network classifier, effectively suppressing the impact of adversarial perturbations on the network. This design allows the network to maintain a high accuracy while robustly resisting malicious attacks on adversarial examples. The experimental results demonstrate that our proposed method outperforms the existing literature, reaching optimal levels on the MIT-BIH and CPSC2018 datasets. This substantiates the effectiveness of our approach in defending against adversarial attacks.

Author Contributions

Conceptualization, X.C. and Y.S.; methodology, X.C.; software, X.C.; validation, X.C. and J.F.; formal analysis, W.Y. and Z.Z.; resources, Y.S.; data curation, W.Y., J.F. and Z.Z.; writing—original draft preparation, X.C. and Y.S.; writing—review and editing, X.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Natural Science Foundation of Guangdong Province (2023A1515011302), the Guangdong Key Disciplines Project (2022ZDJS140), and the Featured Innovation Projects of the Guangdong Universities (2022KTSCX189).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used in this study is publicly available on Kaggle at https://www.kaggle.com/datasets/shayanfazeli/heartbeat and Icbeb at http://2018.icbeb.org/ (accessed on 30 April 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Xiao, Q.; Lee, K.; Mokhtar, S.A.; Ismail, I.; Pauzi, A.L.b.M.; Zhang, Q.; Lim, P.Y. Deep Learning-Based ECG Arrhythmia Classification: A Systematic Review. Appl. Sci. 2023, 13, 4964. [Google Scholar] [CrossRef]
  2. Wu, M.; Lu, Y.; Yang, W.; Wong, S.Y. A study on arrhythmia via ECG signal classification using the convolutional neural network. Front. Comput. Neurosci. 2021, 14, 564015. [Google Scholar] [CrossRef] [PubMed]
  3. Oh, S.L.; Ng, E.Y.; San Tan, R.; Acharya, U.R. Automated diagnosis of arrhythmia using combination of CNN and LSTM techniques with variable length heart beats. Comput. Biol. Med. 2018, 102, 278–287. [Google Scholar] [CrossRef]
  4. Wang, Y.; Yang, G.; Li, S.; Li, Y.; He, L.; Liu, D. Arrhythmia classification algorithm based on multi-head self-attention mechanism. Biomed. Signal Process. Control 2023, 79, 104206. [Google Scholar] [CrossRef]
  5. Kim, Y.K.; Lee, M.; Song, H.S.; Lee, S.-W. Automatic cardiac arrhythmia classification using residual network combined with long short-term memory. IEEE Trans. Instrum. Meas. 2022, 71, 4005817. [Google Scholar] [CrossRef]
  6. Kumar, S.; Mallik, A.; Kumar, A.; Del Ser, J.; Yang, G. Fuzz-ClustNet: Coupled fuzzy clustering and deep neural networks for Arrhythmia detection from ECG signals. Comput. Biol. Med. 2023, 153, 106511. [Google Scholar] [CrossRef]
  7. Zeng, Y.; Lv, H.; Jiang, M.; Zhang, J.; Xia, L.; Wang, Y.; Wang, Z. Deep arrhythmia classification based on SENet and lightweight context transform. Math. Biosci. Eng. 2023, 20, 1–17. [Google Scholar] [CrossRef]
  8. Wu, D.; Xu, J.; Fang, W.; Zhang, Y.; Yang, L.; Xu, X.; Luo, H.; Yu, X. Adversarial attacks and defenses in physiological computing: A systematic review. Natl. Sci. Open 2023, 2, 20220023. [Google Scholar] [CrossRef]
  9. Xu, H.; Ma, Y.; Liu, H.-C.; Deb, D.; Liu, H.; Tang, J.-L.; Jain, A.K. Adversarial attacks and defenses in images, graphs and text: A review. Int. J. Autom. Comput. 2020, 17, 151–178. [Google Scholar] [CrossRef]
  10. Karimian, N.; Woodard, D.; Forte, D. ECG biometric: Spoofing and countermeasures. IEEE Trans. Biom. Behav. Identity Sci. 2020, 2, 257–270. [Google Scholar] [CrossRef]
  11. Chen, H.; Huang, C.; Huang, Q.; Zhang, Q.; Wang, W. Ecgadv: Generating adversarial electrocardiogram to misguide arrhythmia classification system. Proc. AAAI Conf. Artif. Intell. 2020, 34, 3446–3453. [Google Scholar] [CrossRef]
  12. Garg, A.; Karimian, N. ECG biometric spoofing using adversarial machine learning. In Proceedings of the 2021 IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 10–12 January 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–5. [Google Scholar]
  13. Han, X.; Hu, Y.; Foschini, L.; Chinitz, L.; Jankelson, L.; Ranganath, R. Deep learning models for electrocardiograms are susceptible to adversarial attack. Nat. Med. 2020, 26, 360–363. [Google Scholar] [CrossRef]
  14. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. arXiv 2014, arXiv:1412.6572. [Google Scholar]
  15. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards deep learning models resistant to adversarial attacks. arXiv 2017, arXiv:1706.06083. [Google Scholar]
  16. Carlini, N.; Wagner, D. Towards evaluating the robustness of neural networks. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 39–57. [Google Scholar]
  17. Wiedeman, C.; Wang, G. Decorrelative Network Architecture for Robust Electrocardiogram Classification. arXiv 2022, arXiv:2207.09031. [Google Scholar]
  18. Jeong, H.; Son, J.; Kim, H.; Kang, K. Defensive Adversarial Training for Enhancing Robustness of ECG based User Identification. In Proceedings of the 2022 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Las Vegas, NV, USA, 6–8 December 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 3362–3369. [Google Scholar]
  19. Ma, L.; Liang, L. A regularization method to improve adversarial robustness of neural networks for ECG signal classification. Comput. Biol. Med. 2022, 144, 105345. [Google Scholar] [CrossRef]
  20. Shao, J.; Geng, S.; Fu, Z.; Xu, W.; Liu, T.; Hong, S. CardioDefense: Defending against adversarial attack in ECG classification with adversarial distillation training. Biomed. Signal Process. Control 2024, 91, 105922. [Google Scholar] [CrossRef]
  21. Fawzi, A.; Fawzi, O.; Frossard, P. Analysis of classifiers’ robustness to adversarial perturbations. Mach. Learn. 2018, 107, 481–508. [Google Scholar] [CrossRef]
  22. Huang, H.; Wang, Y.; Erfani, S.; Gu, Q.; Bailey, J.; Ma, X. Exploring architectural ingredients of adversarially robust deep neural networks. Adv. Neural Inf. Process. Syst. 2021, 34, 5545–5559. [Google Scholar]
  23. Peng, S.; Xu, W.; Cornelius, C.; Hull, M.; Li, K.; Duggal, R.; Phute, M.; Martin, J.; Chau, D.H. Robust principles: Architectural design principles for adversarially robust cnns. arXiv 2023, arXiv:2308.16258. [Google Scholar]
  24. Yan, H.; Zhang, J.; Niu, G.; Feng, J.; Tan, V.; Sugiyama, M. CIFS: Improving Adversarial Robustness of CNNs via Channel-wise Importance-based Feature Selection. In Proceedings of the 38th International Conference on Machine Learning, Virtual, 18–24 July 2021; Marina, M., Tong, Z., Eds.; PMLR: Proceedings of Machine Learning Research. Volume 139, pp. 11693–11703. [Google Scholar]
  25. Zhou, S.; Schoellig, A.P. An analysis of the expressiveness of deep neural network architectures based on their lipschitz constants. arXiv 2019, arXiv:1912.11511. [Google Scholar]
  26. Kachuee, M.; Fazeli, S.; Sarrafzadeh, M. Ecg heartbeat classification: A deep transferable representation. In Proceedings of the 2018 IEEE International Conference on Healthcare Informatics (ICHI), New York, NY, USA, 4–7 June 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 443–444. [Google Scholar]
  27. ECG Heartbeat Categorization Dataset. Available online: https://www.kaggle.com/shayanfazeli/heartbeat (accessed on 30 April 2024).
  28. Liu, F.; Liu, C.; Zhao, L.; Zhang, X.; Wu, X.; Xu, X.; Liu, Y.; Ma, C.; Wei, S.; He, Z. An open access database for evaluating the algorithms of electrocardiogram rhythm and morphology abnormality detection. J. Med. Imaging Health Inform. 2018, 8, 1368–1373. [Google Scholar] [CrossRef]
  29. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  30. Bai, Y.; Zeng, Y.; Jiang, Y.; Xia, S.-T.; Ma, X.; Wang, Y. Improving adversarial robustness via channel-wise activation suppressing. arXiv 2021, arXiv:2103.08307. [Google Scholar]
  31. Zhang, B.; Cai, T.; Lu, Z.; He, D.; Wang, L. Towards certifying robustness using neural networks with l-dist neurons. arXiv 2021, arXiv:2102.05363. [Google Scholar]
  32. Wang, Y.; Zou, D.; Yi, J.; Bailey, J.; Ma, X.; Gu, Q. Improving adversarial robustness requires revisiting misclassified examples. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  33. Zhang, H.; Yu, Y.; Jiao, J.; Xing, E.; El Ghaoui, L.; Jordan, M. Theoretically principled trade-off between robustness and accuracy. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 7472–7482. [Google Scholar]
  34. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  35. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
Figure 1. Overall flow of the proposed methodology.
Figure 1. Overall flow of the proposed methodology.
Sensors 24 02954 g001
Figure 2. CASLCNet framework.
Figure 2. CASLCNet framework.
Sensors 24 02954 g002
Figure 3. Dimension reduction diagram.
Figure 3. Dimension reduction diagram.
Sensors 24 02954 g003
Figure 4. Illustration of channel-wise activation frequency and magnitude using CASLCNet and Resnet18 on MIT-BIH datasets. (a) Channel-wise activation magnitude of Resnet18; (b) channel-wise activation magnitude of CASLCNet; (c) channel-wise activation frequency of Resnet18; and (d) channel-wise activation frequency of CASLCNet.
Figure 4. Illustration of channel-wise activation frequency and magnitude using CASLCNet and Resnet18 on MIT-BIH datasets. (a) Channel-wise activation magnitude of Resnet18; (b) channel-wise activation magnitude of CASLCNet; (c) channel-wise activation frequency of Resnet18; and (d) channel-wise activation frequency of CASLCNet.
Sensors 24 02954 g004
Figure 5. Illustration of channel-wise activation frequency and magnitude using CASLCNet and Resnet18 on CPSC2018 datasets. (a) Channel-wise activation magnitude of Resnet18; (b) channel-wise activation magnitude of CASLCNet; (c) channel-wise activation frequency of Resnet18; and (d) channel-wise activation frequency of CASLCNet.
Figure 5. Illustration of channel-wise activation frequency and magnitude using CASLCNet and Resnet18 on CPSC2018 datasets. (a) Channel-wise activation magnitude of Resnet18; (b) channel-wise activation magnitude of CASLCNet; (c) channel-wise activation frequency of Resnet18; and (d) channel-wise activation frequency of CASLCNet.
Sensors 24 02954 g005
Figure 6. CASLCNet network loss function and accuracy display in MIT-BIH and CPSC2018 dataset. (a) CASLCNet indicator in MIT-BIH dataset; and (b) CASLCNet indicator in CPSC2018 dataset.
Figure 6. CASLCNet network loss function and accuracy display in MIT-BIH and CPSC2018 dataset. (a) CASLCNet indicator in MIT-BIH dataset; and (b) CASLCNet indicator in CPSC2018 dataset.
Sensors 24 02954 g006
Figure 7. Accuracy and F1 scores of proposed method with Jacob and SNR under different adversarial attacks in MIT-BIH dataset. (a) Accuracy under different adversarial perturbations in PGD adversarial attacks; (b) F1 score under different adversarial perturbations in PGD adversarial attacks; (c) accuracy under different adversarial perturbations in SAP adversarial attacks; and (d) F1 score under different adversarial perturbations in SAP adversarial attacks.
Figure 7. Accuracy and F1 scores of proposed method with Jacob and SNR under different adversarial attacks in MIT-BIH dataset. (a) Accuracy under different adversarial perturbations in PGD adversarial attacks; (b) F1 score under different adversarial perturbations in PGD adversarial attacks; (c) accuracy under different adversarial perturbations in SAP adversarial attacks; and (d) F1 score under different adversarial perturbations in SAP adversarial attacks.
Sensors 24 02954 g007
Figure 8. Accuracy and F1 scores of proposed method with Jacob and SNR under different adversarial attacks in CPSC2018 dataset. (a) Accuracy under different adversarial perturbations in PGD adversarial attacks; (b) F1 score under different adversarial perturbations in PGD adversarial attacks; (c) accuracy under different adversarial perturbations in SAP adversarial attacks; and (d) F1 score under different adversarial perturbations in SAP adversarial attacks.
Figure 8. Accuracy and F1 scores of proposed method with Jacob and SNR under different adversarial attacks in CPSC2018 dataset. (a) Accuracy under different adversarial perturbations in PGD adversarial attacks; (b) F1 score under different adversarial perturbations in PGD adversarial attacks; (c) accuracy under different adversarial perturbations in SAP adversarial attacks; and (d) F1 score under different adversarial perturbations in SAP adversarial attacks.
Sensors 24 02954 g008
Table 1. Dataset details.
Table 1. Dataset details.
DatasetSampling Rates (Hz)ECG TypeNumber of Heartbeats
Training SetTest Set
MIT-BIH125N72,47118,118
S2223556
V57881448
F641162
Q64311608
CPSC2018500Normal86850
AF92650
I-AVB63650
LBBB12950
RBBB148350
PAC48250
PVC55750
STD73450
STE13550
Table 2. Detailed settings of CASLCNet.
Table 2. Detailed settings of CASLCNet.
Layer NameMIT-BIHCPSC2018
Output SizeModule Parameters
( Kernel   Size   k s ,   Dimension   d ,   Stride   s t )
Output SizeModule Parameters
( Kernel   Size   k s ,   Dimension   d ,   Stride   s t )
1D convolution layer64 × 94 k s = 11 ,   s t = 2 ,   d = 6464 × 16,896 k s = 11 ,   s t = 2 ,   d = 64
Group Normalization64 × 94[64,64]64 × 16,896[64,64]
Leaky ReLU64 × 94-64 × 16,896-
Residual layer 1128 × 24 k s = 3 , d = 128 k s = 3 , d = 128 × 2 ,   s t = 2128 × 4224 k s = 3 , d = 128 k s = 3 , d = 128 × 2 ,   s t = 2
Residual layer 2256 × 6 k s = 3 , d = 256 k s = 3 , d = 256 × 2 ,   s t = 2256 × 1056 k s = 3 , d = 256 k s = 3 , d = 256 × 2 ,   s t = 2
Residual layer 3512 × 2 k s = 3 , d = 512 k s = 3 , d = 512 × 2 ,   s t = 2512 × 264 k s = 3 , d = 512 k s = 3 , d = 512 × 2 ,   s t = 2
Residual with CAS layer512 × 2 k s = 3 , d = 512 k s = 3 , d = 512 × 2 ,   s t = 1512 × 132 k s = 3 , d = 512 k s = 3 , d = 512 × 2 ,   s t = 2
Downscaling (average pooling)512 s t = 2512 k s = 3072 ,   s t = 234
distance network5 512 , 512 , 512 , 512 , 512 , 5 9 512 , 512 , 512 , 512 , 512 , 5
Table 3. Impact of hyperparameter α on model adversarial robustness (%).
Table 3. Impact of hyperparameter α on model adversarial robustness (%).
DatasetHyperparameter
α
PGDSAP
A C C r o b u s t F 1 r o b u s t A C C r o b u s t F 1 r o b u s t
MIT-BIH060.6856.6363.1259.37
187.3387.4888.1588.54
291.0391.2291.9092.07
385.5085.7286.0986.33
486.7786.9387.2187.45
CPSC2018080.6180.4681.8481.70
179.1878.8583.5283.27
283.0182.6485.3485.07
380.0179.9981.6281.66
478.7878.3280.7580.35
Table 4. Performance metrics of CASCLC-Net in MIT-BIH and CPSC2018 dataset under different adversarial attack scenarios (%).
Table 4. Performance metrics of CASCLC-Net in MIT-BIH and CPSC2018 dataset under different adversarial attack scenarios (%).
DatasetMetrics-White NoiseFGSMMI-FGSMC&W
MIT-BIHAccuracy92.4491.5889.5688.4386.25
F192.4892.0990.3488.4986.72
CPSC2018Accuracy84.8984.8986.6761.3360.89
F184.6185.0786.4662.3457.38
Table 5. Performance metrics of CASCLC-Net in MIT-BIH dataset under PGD and SAP adversarial attack scenarios (%).
Table 5. Performance metrics of CASCLC-Net in MIT-BIH dataset under PGD and SAP adversarial attack scenarios (%).
AttackMetrics00.010.030.050.10.20.3
PGDAccuracy92.44 92.1891.1990.3985.2653.2823.24
F192.4892.2491.3790.7085.9755.3023.82
SAPAccuracy92.44 92.3091.9691.6989.6383.5878.17
F192.4892.3892.1392.1290.0484.0678.71
Table 6. Performance metrics of CASCLC-Net in CPSC2018 dataset under PGD and SAP adversarial attack scenarios (%).
Table 6. Performance metrics of CASCLC-Net in CPSC2018 dataset under PGD and SAP adversarial attack scenarios (%).
AttackMetrics00.0010.0030.0050.0070.010.030.050.1
PGDAccuracy84.8984.8985.3385.5684.8984.8984.0082.0074.67
F184.6184.6185.0785.3084.6384.6383.7581.6873.76
SAPAccuracy84.8984.8984.8984.8985.1185.1185.3386.0086.44
F184.6184.6184.6184.6184.8584.8585.0985.7786.20
Table 7. Results of ablation experiments (%).
Table 7. Results of ablation experiments (%).
DatasetMethodBackBoneCAS Block Distance NetPGD SAP
A C C r o b u s t F 1 r o b u s t A C C r o b u s t F 1 r o b u s t
MIT-BIH1ResNet18××68.8183.3986.6084.83
2ResNet18×82.9586.1888.2388.33
3ResNet18×85.6786.5184.7085.08
4ResNet1891.0391.2291.9092.07
CPSC20181ResNet18××31.8531.8072.5771.92
2ResNet18×76.0775.1281.4780.56
3ResNet18×50.2851.1681.4481.22
4ResNet1883.0182.6485.3485.07
Table 8. Comparison of results of different adversarial training methods (%).
Table 8. Comparison of results of different adversarial training methods (%).
DatasetMethodPGDSAP
A C C r o b u s t F 1 r o b u s t A C C r o b u s t F 1 r o b u s t
MIT-BIHAdversarial Training91.0291.1192.0692.50
Trades88.0488.1889.3489.49
MART91.0391.2291.9092.07
CPSC2018Adversarial Training52.3451.9779.0578.76
Trades74.5873.8283.2382.91
MART83.0182.6485.3485.07
Table 9. Contrast results with different methods in MIT-BIH dataset (%).
Table 9. Contrast results with different methods in MIT-BIH dataset (%).
MethodPGDSAP
A C C r o b u s t F 1 r o b u s t A C C r o b u s t F 1 r o b u s t
VGG1935.8536.9463.6964.31
VGG19+MART82.2182.3080.0483.91
ResNet1856.7957.6665.7865.97
ResNet18+ MART68.8183.3986.6084.83
DenseNet71.1869.2079.9178.27
DenseNet + MART83.5183.6185.3185.39
CASLCNet53.0451.3565.1864.47
CASLCNet + MART91.0391.2291.9092.07
Table 10. Contrast results with different methods in CPSC2018 dataset (%).
Table 10. Contrast results with different methods in CPSC2018 dataset (%).
MethodPGDSAP
A C C r o b u s t F 1 r o b u s t A C C r o b u s t F 1 r o b u s t
VGG1921.9523.1966.1465.19
VGG19+MART30.1730.2370.1067.26
ResNet1818.4218.4339.4738.76
ResNet18+ MART31.8531.8072.5771.92
DenseNet22.2221.8974.0073.48
DenseNet + MART44.5043.9876.7975.70
CASLCNet46.7645.9984.7684.74
CASLCNet + MART83.0182.6485.3485.07
Table 11. Comparative experimental results with other methods (%).
Table 11. Comparative experimental results with other methods (%).
AuthorMethodDatasetPGDSAP
A C C r o b u s t F 1 r o b u s t A C C r o b u s t F 1 r o b u s t
Ma et al. [19]JabobMIT-BIH79.9479.8879.8882.93
CPSC201848.9149.1880.6070.12
SNRMIT-BIH88.6488.6388.6391.08
CPSC201846.9946.6781.2680.98
OursCASLCNetMIT-BIH91.0391.2291.9092.07
CPSC201883.0182.6485.3485.07
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, X.; Si, Y.; Zhang, Z.; Yang, W.; Feng, J. Improving Adversarial Robustness of ECG Classification Based on Lipschitz Constraints and Channel Activation Suppression. Sensors 2024, 24, 2954. https://doi.org/10.3390/s24092954

AMA Style

Chen X, Si Y, Zhang Z, Yang W, Feng J. Improving Adversarial Robustness of ECG Classification Based on Lipschitz Constraints and Channel Activation Suppression. Sensors. 2024; 24(9):2954. https://doi.org/10.3390/s24092954

Chicago/Turabian Style

Chen, Xin, Yujuan Si, Zhanyuan Zhang, Wenke Yang, and Jianchao Feng. 2024. "Improving Adversarial Robustness of ECG Classification Based on Lipschitz Constraints and Channel Activation Suppression" Sensors 24, no. 9: 2954. https://doi.org/10.3390/s24092954

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop