Push-Pull: Characterizing the Adversarial Robustness for Audio-Visual
Active Speaker Detection
Abstract
Audio-visual active speaker detection (AVASD) is well-developed, and now is an indispensable front-end for several multi-modal applications. However, to the best of our knowledge, the adversarial robustness of AVASD models hasn’t been investigated, not to mention the effective defense against such attacks. In this paper, we are the first to reveal the vulnerability of AVASD models under audio-only, visual-only, and audio-visual adversarial attacks through extensive experiments. What’s more, we also propose a novel audio-visual interaction loss (AVIL) for making attackers difficult to find feasible adversarial examples under an allocated attack budget. The loss aims at pushing the inter-class embeddings to be dispersed, namely non-speech and speech clusters, sufficiently disentangled, and pulling the intra-class embeddings as close as possible to keep them compact. Experimental results show the AVIL outperforms the adversarial training by 33.14 mAP (%) under multi-modal attacks.
Index Terms— Audio-visual active speaker detection, multi-modal adversarial attack, adversarial robustness
1 Introduction
Active Speaker Detection (ASD) seeks to detect who is speaking in a visual scene containing one or more speakers [1, 2]. Recently, audio-visual ASD (AVASD), which integrates audio-visual information by learning the relationship between speech and facial motion, effectively improves the performance of ASD, and AVASD has become more indispensable as a front-end for multi-modal applications. However, to the best of our knowledge, whether the AVASD models are robust against adversarial attacks has not been investigated previously, not to mention the effective defense method against such multi-modal attacks.
Crafting indistinguishable adversarial noise, adding such noise to clean samples to generate adversarial samples, and then manipulating the AI models by such samples, is called adversarial attack [3]. Previous adversarial attacks usually focus on single-modal applications. For visual-modal attacks, Szegedy et al. first propose to attack state-of-the-art image classification models [3] in 2013. For the speech modality, models including automatic speaker verification (ASV) systems [4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18], anti-spoofing models for ASV [19, 20, 21, 22, 23], and automatic speech recognition models [24, 25, 26, 27, 28, 29, 30] are also vulnerable to adversarial attacks. For audio-visual learning, Li et al. [31] studied the audio-visual adversarial robustness of the general sound event detection model but only considered single- or multi-modal attacks under an attack method.
Given that AVASD is now ubiquitously implemented as a front-end for a variety of multi-modal downstream models, the dangerous adversarial noise may manipulate the AVASD front-end to commit errors, which will accumulate and propagate to the downstream applications. Hence it is of high priority that we mitigate the adversarial vulnerability of AVASD and ensure robustness against such attacks. This paper investigates the susceptibility of AVASD models to adversarial attacks and then proposes a novel defense method to improve their robustness. Our contributions are summarized in two folds: 1). To the best of our knowledge, this is the first work to reveal the vulnerability of AVASD models under three kinds of attacks, including audio-only, visual-only, and audio-visual adversarial attacks by extensive experiments. 2). We also propose a novel audio-visual interaction loss (AVIL), which aims at pushing the inter-class embeddings, namely the non-speech and speech clusters, sufficiently disentangled, and pulling the intra-class embeddings as close as possible. Expanding the inter-class dispersion and enhancing the intra-class compactness will make it difficult for attackers to find feasible adversarial samples to go beyond the decision boundary within the allocated attacking budget. The experimental results illustrate that the brand-new audio-visual interaction loss effectively strengthens the invulnerability of AVASD models.

2 Background
2.1 Audio-Visual Active Speaker Detection
The ASD task has been studied using audio, video, or the fusion of both. For audio, the voice activity detector [32, 33] is often used to detect the presence of speech. However, in real-world scenarios, the speech signal from the microphones is easily mixed with overlapping speech and background noise, which will hinder the effectiveness of voice activity detection. The visual part [34, 35] mainly analyzes the face and upper body of a person to determine whether the person is speaking, but the performance is limited due to some non-speech activities, e.g. licking lips, eating, and grinning. The audio-visual processing refers to the combination of audio and visual parts [36, 37], and allows learning across modalities about the relationship between audio speech and facial motions. With valuable support from sizeable datasets, e.g. AVA-Active Speaker, and the AVA Challenge series launched since 2018, a variety of high-performance models for AVASD have emerged recently [38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48]. In real-world user authentication systems, AVASD can be used as a front-end task to assure security verification for speaker verification [49]. For AVASD system, there are four typical cases, such as speech without target speaker, no audible speaker, speaker without speech and speech with the target speaker. Only the speech with target speaker is labeled as speaking. Attackers possibly use some single modal attack methods or combine them to make AVASD produce wrong predictions in the other three cases, which is dangerous. However, people have not yet seen investigations on the adversarial robustness of AVASD model.
2.2 Adversarial Attacks
Adversarial attack is to manipulate a well-trained model to give wrong predictions by an adversarial sample, which is imperceptible by humans, compared with the original (unmodified) counterpart. Mathematically, given a clean sample and the ground-truth label , attack algorithms seek to find a sufficiently small perturbation such that: , where is the adversarial sample that can fool the model to produce the wrong prediction . We can find a suitable by solving the following objective function:
(1) | |||
where denotes the objective function pre-defined by the attackers, which is usually set to maximize the difference between and the model’s final prediction given , is the allowed perturbation budget, and denotes the -norm, which is usually considered to be a -ball or -ball centered at . We evaluate the AVASD models’ vulnerability with -boundary adversarial noise, as it is widely used as a standard evaluation boundary for adversarial robustness [50]. To solve the optimization problem as shown in Equation 1, we choose three widely used attack methods to evaluate the robustness of AVASD models, and the details are summarized below.
Basic Iterative Method (BIM) BIM [51] is a method with iterative updates as follows:
(2) | |||
where starts from the original sample , is the step size, is the number of iterations and the function applies element-wise clipping to make . The perturbed example is the final adversarial example.
Momentum-based Iterative Method (MIM) MIM [52] is an improved version of BIM. MIM introduces a momentum term into the iterative process to avoid BIM falling into local minimum and thus improve the attack performance over BIM.
Projected Gradient Descent (PGD) PGD [50] is also a variant of BIM. PGD randomly initializes the adversarial noise for times and conducts BIM-style attacks to generate candidates of adversarial noise. Finally, the best one out of the candidates with the best attack performance, will be chosen as the final adversarial sample.

3 Methodology
3.1 AVASD Model – TalkNet
We adopt TalkNet [48] for our case study to characterize the adversarial robustness of AVASD. TalkNet is one of the state-of-the-art models for AVASD, which is fully end-to-end. TalkNet takes a sequence of video frames consisting of cropped face sequences and the corresponding audio sequences as inputs. The output probability denotes how likely the person is speaking in the given video frame. TalkNet comprises a feature representation front-end, and a speaker detection back-end classifier, as shown in Fig. 1.(a). The front-end consists of an audio temporal encoder and a video temporal encoder to extract audio embeddings and visual embeddings for the frame. In the back-end, the audio and visual embeddings are aligned via inter-modality cross-attention and then concatenated to obtain the joint audio-visual embeddings and for the frame. Then, a self-attention network is applied after the cross-attention network to model the audio-visual temporal information. Finally, a fully-connected layer with a softmax is implemented to project the output of the self-attention network to a sequence of ASD labels. The predicted label sequence is compared with the ground-truth label sequence by cross-entropy loss ():
(3) |
where and are the ground-truth and predicted score for the frame, and is the total frames for one sample of video data. During training, TalkNet utilizes two additional predict heads for audio embeddings and visual embeddings after the cross-attention module to predict the ASD label sequences, as shown in the part of the speaker detection back-end in Fig. 1.(a). The additional outputs here are used to calculate the weighted loss, and the final training loss is shown as follows:
(4) |
where and denote the losses of audio-only and visual-only prediction head, respectively. The coefficient 0.4 is referred from the TalkNet [48]. During inference, only the prediction head after self-attention will be utilized. For further details of the above setting, please refer to the TalkNet paper [48].
3.2 Multi-Modal Attacks
Let be an audio input, be the visual input and be the corresponding ground-truth label for the multisensory input: {, }. We divide the audio-visual adversarial attack into three categories: the audio-only attack generates the audio adversarial example , the visual-only attack generates the visual adversarial example and the audio-visual attack generates multi-modal adversarial examples: , . To force a well-trained audio-visual model to make wrong predictions with corresponding perturbations being as imperceptible as possible, the objective function for multi-modal attacks is as follows:
(5) | |||
where , , is the objective function to make the outputs of the audio-visual model as different as possible to , is the -norm, and and are audio and visual perturbation budgets. In the case of an audio-only attack, the perturbation budget is equal to 0, and in the case of a visual-only attack, the perturbation budget is equal to 0. In the case of audio-visual attacks, both audio and visual inputs will be perturbed. Fig. 1(b) illustrates the audio-visual adversarial attack framework. Different strategies to search for , which consists of and , result in different adversarial attack methods. Note that our multi-modal attack is jointly optimized on audio-visual modality instead of optimizing independently. This paper adopts three famous attack methods for their effective attacking performance and affordable execution time based on our resources: BIM, MIM, and PGD.
We also set up two attack scenarios: training-aware attack and inference-aware attack scenarios. In both kinds of attack scenarios, the attackers have full access to the model internals, including model architectures, parameters, and gradients. Besides, the inference-aware attackers know exactly the inference procedure of the AVASD model. In other words, they know the prediction head adopted for inference is the audio-visual head after the self-attention as shown in Fig. 1(a), and then they will conduct adversarial attacks based on the loss as shown in Equation 3. The inference-aware attack scenario is more practical, as it relies on the real inference procedure. For the training-aware attackers, they even know the training loss of the AVASD model, and they will conduct adversarial attacks by Equation 4. Training-aware attacks can craft even more dangerous attacks as they adopt all three prediction heads to find the adversarial perturbation. Unless specified otherwise, all the experiments are conducted under the training-aware attack scenario as it is more dangerous. We also perform the experiments for the inference-aware scenario and it shows the same trend. We show comparison results between training-aware and inference-aware attacks in Section 4.4.
3.3 Audio-Visual Interaction Loss
In this section, we first introduce the proposed audio-visual interaction loss (AVIL) and the implementation details. Then we will present the rationale of the proposed method.
Implementation Procedure of AVIL. Suppose we have frames for one batch, and let and be the speech and non-speech frame numbers, respectively. Let and denote the index sets for speech and non-speech. We can get the four centers as below:
(6) | |||||
where , , , denote the centers for audio speech embeddings, audio non-speech embeddings, visual speech embeddings, visual non-speech embeddings, respectively. The centers are denoted with bold borders as shown in Fig. 2.
Then we can define the four audio-visual interaction losses:
- •
-
•
Intra-modality intra-class dissimilarity (Fig. 2.(b)):
(8) -
•
Inter-modality intra-class dissimilarity (center-based) as shown Fig. 2.(c):
(9) -
•
Inter-modality intra-class distance (sample-based) as shown in Fig. 2.(d):
(10) where and denote the speech embeddings for the audio and visual modalities, respectively. When , and denote the non-speech embeddings.
To alleviate the adversarial noise, the four above losses mentioned above are adapted in the training process of the AVASD model and the final objective function is formulated as:
(11) |
where denotes the hyperparameter. Note that if , will reduce to , the training loss of the original TalkNet. For training the model with AVIL, we simply set , for simplicity. And also, we just want to show the effectiveness of the four audio-visual interaction losses, rather than exhaust the hyperparameter settings and trickly improve the defense performance.
Rationale of AVIL. The adversarial attacks threaten the AVASD model by maximizing the loss functions, e.g. Equation 3 and Equation 4, and then will urge the output far away from its original decision region [3, 53]. For example, after the adversarial attack, the output for a speech frame will go away from the right region, namely “speech”, and become non-speech. As a result, it is reasonable that high inter-class dispersion and intra-class compactness will boost models’ invulnerability, as it will make it hard for the attackers to find feasible adversarial perturbations within a given budget to push the genuine samples to pass through the decision boundary.
Minimizing will equip the model with better discrimination capacity between speech and non-speech embeddings, resulting in higher inter-class difference from the models’ perspective. Maximizing and minimizing will force the model to render more compact intra-class features. Incorporating in the training process, we can simultaneously urge the model to learn both discriminative inter-class features, and compact intra-class features, leading the model less susceptible to adversarial perturbations. As shown in Table 1, the four losses achieve the goal of significantly improving the robustness of the models.

4 Experiment
4.1 Experimental setup
We use TalkNet [48] to investigate the adversarial robustness of AVASD and verify the effectiveness of our proposed method to alleviate adversarial attacks. We reproduce and modify the TalkNet based on the official TalkNet GitHub repository to attack and defend. To conduct gradient-based adversarial attacks, including BIM, MIM, and PGD, we revise the feature extraction and data augmentation steps using the PyTorch library to make the entire model pipeline differentiable. For the dataset, we use the AVA Active Speaker dataset [1], which contains 29,723 video samples for training. Since the ground-truth labels for the testing set are not available to the public, we use the validation set with 8,015 samples as the evaluation set. The lengths of videos range from 1 to 10 seconds and their facial tracks are also provided. We follow the official evaluation plan and evaluate performance with mean average precision (mAP) [1]. The revised TalkNet achieves 92.58% mAP on the evaluation set, which is slightly higher than the original paper [48]. As the adversarial attack is time- and resource-consuming, we randomly selected 450 genuine samples (225 speaking and 225 non-speaking) with the correct predictions to conduct adversarial attacks. The imperceptible attack budget of audio and visual modality has a very large numerical gap, we introduced to represent the attack budget for easier explanation. The relationships between and attack budget of two modalities are and , respectively.
4.2 The Model Vulnerability under Multi-modality Attacks
Fig. 3 illustrates the attack performance on AVASD under both single-modality and audio-visual attacks, with three attack algorithms in both white-box and black-box attack scenarios. The blue line is the baseline, where the genuine samples are fed directly into the TalkNet model without any attacks. To investigate the vulnerability of AVASD under both black-box and white-box scenarios, we also trained two models, ncTalNet and specTalkNet. ncTalkNet represents the TalkNet model without the cross attention module, as shown in Fig. 1 (a). specTalkNet denotes the TalkNet by replacing the audio features with linear spectrograms rather than MFCCs adopted by the original TalkNet.
White-box and Black-box Attackers. In Fig. 3 (a), there are three settings. The TalkNet-TalkNet denotes the white-box scenario, that is, both the model for generating adversarial samples and the target model are TalkNet. The ncTalkNet-TalkNet and specTalkNet-TalkNet are the black-box scenarios, in which the substitute models for generating adversarial samples are ncTalkNet specTalkNet, respectively, and the target model is TalkNet. White-box attackers achieve effective attack performance by degrading the mAP of the TalkNet to a large scale, while black-box attackers can barely manipulate the TalkNet. TalkNet-TalkNet degrades the mAP from 100% to 65.2%, when under multi-modal attack with PGD method. But in the same situation, black-box attackers are almost ineffective.
Single-modal and Multi-modal Attacks. We take TalkNet-TalkNet to further evaluate the attack performance under single-modal, and multi-modal with the PGD method in Fig. 3 (b). When , the audio-only, visual-only attack, and multi-modal attacks can degrade the model’s mAP to 99%, 77% and 65.2%. We can observe the same phenomenon in other settings. As a result, we can derive the following three results: (1) Audio-only attacks can barely influence the mAP of AVASD. (2) Visual-only attacks achieve more effective attack performance than audio-only attacks. One possible reason is that for one video data sample, there are far fewer audio samples than the pixels [3]. (3) Multi-modal attacks are always more dangerous than single-modal attacks.
Different Attack Algorithms with Different Attack Budgets. We take TalkNet-TalkNet to further evaluate the multi-modal attack performance under BIM, MIM, and PGD attacks. From Fig. 3 (c), we have the following observations: (1) As the attack budget increases, the mAP of TalkNet has a significant decrease trend. (2) All three attack methods can effectively degrade the AVASD model. In the following experiments to evaluate the defense performance, we only consider multi-modal attacks in the white-box scenarios, since it is the most dangerous one, and attack budgets are set as .
The Imperceptibility of Multi-Modal Attacks. We conduct the XAB test to verify that the adversarial noise generated by multi-modal attacks is both visually and acoustically imperceptible. The XAB test is a standard test to evaluate the detectability between two sensory stimulus choices. We randomly select 4 adversarial-genuine pairs (2 speech and 2 non-speech pairs) for each of the three attack algorithms with , resulting in 12 pairs of randomly selected adversarial-genuine pairs (i.e., A and B). One reference data (i.e., X) is chosen from A and B. A, B, and X are shown to the volunteers. The volunteers should select the more similar data to X, from A and B. We hire five volunteers to join in the XAB test. The classification accuracy for the XAB test is 53.33%, which is a nearly random guess, leading to the conclusion that the adversarial samples are difficult to be distinguished from genuine samples. The XAB test samples will be shown here 111https://xjchengit.github.io/Push-Pull/index.html.
Model | Adversarial training [54] | Clean mAP (%) | Different attack methods with | |||||||||
BIM | MIM | PGD | ||||||||||
A (ECR) | V (ECR) | mAP (%) | A (ECR) | V (ECR) | mAP (%) | A (ECR) | V (ECR) | mAP (%) | ||||
(A) | ✗ | 92.58 | 0.1658 | 0.3140 | 49.53 | 0.1683 | 0.3170 | 49.30 | 0.1715 | 0.3236 | 47.79 | |
(B1) | BIM | 92.15 | 0.2759 | 0.2644 | 62.7 | 0.2778 | 0.2663 | 59.26 | 0.2937 | 0.2772 | 60.01 | |
(B2) | MIM | 91.34 | 0.3052 | 0.2108 | 54.66 | 0.3073 | 0.2133 | 52.18 | 0.3030 | 0.2118 | 54.23 | |
(B3) | PGD | 91.68 | 0.2728 | 0.1846 | 58.29 | 0.2783 | 0.1893 | 58.3 | 0.2840 | 0.1938 | 56.06 | |
(C1) | ✗ | 92.09 | 0.1407 | 0.2603 | 82.96 | 0.1382 | 0.2618 | 81.32 | 0.1379 | 0.2677 | 80.98 | |
(C2) | ✗ | 92.05 | 0.1451 | 0.1444 | 92.65 | 0.1496 | 0.1481 | 90.69 | 0.1501 | 0.1509 | 88.93 | |
(C3) | ✗ | 92.16 | 0.1575 | 0.3264 | 74.98 | 0.1602 | 0.3289 | 76.25 | 0.1590 | 0.3343 | 73.97 | |
(C4) | ✗ | 91.28 | 0.1065 | 0.2262 | 83.82 | 0.1154 | 0.2322 | 79.82 | 0.1158 | 0.2351 | 78.72 | |
(D1) | ✗ | 92.46 | 0.2182 | 0.4672 | 66.91 | 0.2149 | 0.4656 | 67.89 | 0.2317 | 0.4910 | 64.11 | |
(D2) | ✗ | 92.20 | 0.2218 | 0.4102 | 48.16 | 0.2190 | 0.4134 | 47.92 | 0.2239 | 0.4228 | 49.27 | |
(D3) | ✗ | 91.81 | 0.0820 | 0.2194 | 93.86 | 0.0834 | 0.2337 | 93.34 | 0.0811 | 0.2313 | 93.15 | |
(D4) | ✗ | 92.27 | 0.1525 | 0.3094 | 57.02 | 0.1500 | 0.3029 | 63.36 | 0.1549 | 0.3135 | 61.54 | |
(D5) | ✗ | 91.93 | 0.0936 | 0.1583 | 68.12 | 0.0962 | 0.1612 | 66.28 | 0.0992 | 0.1667 | 67.75 | |
(D6) | ✗ | 91.70 | 0.0782 | 0.2128 | 91.79 | 0.0771 | 0.2135 | 92.48 | 0.0785 | 0.2172 | 91.01 | |
(E1) | BIM | 90.63 | 0.0989 | 0.1007 | 97.85 | 0.1006 | 0.1011 | 97.6 | 0.0955 | 0.1040 | 97.47 | |
(E2) | MIM | 91.70 | 0.0344 | 0.0676 | 99.99 | 0.0341 | 0.0669 | 99.98 | 0.0355 | 0.0696 | 99.97 | |
(E3) | PGD | 91.88 | 0.0470 | 0.1001 | 97.68 | 0.0446 | 0.0966 | 97.47 | 0.0423 | 0.0953 | 98.67 |
4.3 Audio-Visual Interaction Loss
Table 1 shows the defense performance of different methods under three attack algorithms. (A) denotes the original TalkNet model. (B1)-(B3) are the baselines, which represent models trained using adversarial training [54], and the adversarial examples are generated by BIM, MIM, and PGD, respectively. Adversarial training is conducted by injecting the adversarial data into the training set and thus alleviating the adversarial invulnerability. (C1)-(C4) denote the model trained by incorporating only one of the four losses in Section 3.3. We exhaust the pairwise permutation of the four losses to train AVASD models, which are shown in (D1)-(D6). We select the model with the best defense performance from (D1)-(D6), namely (D3), and combine it with adversarial training to see whether the AVIL can complement adversarial training, and the models are shown in (E1)-(E3). The clean mAP(%) column is the mAP performance testing on the entire evaluation set without adversarial attacks. In order to conduct fair comparison, we get the data with correct prediction for model (A)-(E3), and do intersection of such data to get the testing data.
To show the effectiveness of defense methods against adversarial noise, we also set up another evaluation metric, the embedding change ratio (ECR) for audio embedding and visual embedding after the cross-attention as shown in Fig. 1. (a). Take the audio ECR for example, it is calculated by: where and are the adversarial and genuine embeddings, is the number of total frames. ECR measures the embedding change ratio before and after adversarial attacks. A (ECR) and V (ECR) denote the ECR of the audio and visual parts, respectively. The lower the ECR is, the less effect introduced by the attack algorithms and the better the defense performance will be.
Baselines. From (A), the original TalkNet model performs well on the AVASD task with 92.58% mAP for the clean samples, as shown in Table 1. However, the multi-modality attacks seriously degrade the mAP of (A). It can only get 49.53%, 49.40%, and 47.79% mAP under BIM, MIM, and PGD attack algorithms respectively. According to the (B1)-(B3) of Table 1, adversarial training does improve the robustness of the AVASD model. Using the BIM attack algorithm to generate adversarial examples and then conducting adversarial training achieves the best defense performance compared with the other two attack algorithms, resulting in 13.17%, 9.96%, and 12.22% absolute improvement of mAP under BIM, MIM, and PGD attacks, respectively. In terms of ECR, adversarial training effectively reduce V(ECR), yet significantly increase A(ECR).
Using One AVIL. As shown in (C1)-(C4), using any one of the four AVILs improves the mAP to a large scale, resulting in better defense performance the (B1)-(B3), the adversarial training using BIM, MIM, and PGD. (C2), using leads to the best performance. It seems that when only introducing one loss, maximizing the intra-modality and intra-class similarity is the best choice to tackle the adversarial noise. Regarding the ECR, (C1)-(C4) help reduce A(ECR) and V(ECR) in most of the settings.
Pairwise Permutations of AVILs. From (C1)-(D6) in Table 1, we have the following observations and analysis: (1) Adopting pairwise permutations of AVILs perform better than adversarial training in most of the settings from both mAP and ECR perspectives. (2) Combining and even improves the mAP to 93.86%, 93.34%, and 93.15% for BIM, MIM, and PGD respectively. It is the most robust combination. A possible reason is that enlarging inter-class dispersion (minimizing ) and maximizing intra-class similarity (maximizing ) at the same time will result in the model with the best robustness.
Integrating AVIL with Adversarial Training. (E1)-(E3) show that combining AVIL with adversarial training can leverage their complementary to improve the adversarial robustness. Furthermore, integrating AVIL with MIM-based adversarial training can improve the mAP to over 99% under BIM, MIM, and PGD attacks.

4.4 Training-aware and Inference-aware Attacks
Fig. 4. compares the performance of 8 AVASD models in Table 1 under the training-aware and inference-aware scenarios with BIM, MIM, PGD attack method. We also have the same evaluation set as Table 1. The legend show the two scenarios with three attack method. For instance, the legend “ (BIM)” denotes using the to craft the adversarial samples with BIM attack method. The and are defined in equations 4 and 3. We have the following observations and analysis: (1) In (A) groups, we can see that the performance of the AVASD model has also dropped significantly in the inference-aware scenario, but the inference-aware attacks are less dangerous compared with training-aware attacks. (2) From (B1)-(B3), adversarial training does alleviate the adversarial noise with the same trend in the training-aware attack scenario. (3) Compare (D3) with (B1)-(B3), the AVIL performs better in improving the adversarial robustness than adversarial training. (4) From (E1)-(E3), we can see that AVIL can complement adversarial training. To sum up, we can conclude that the inference-aware attack scenario has the same trend as the training-aware attack scenario.
5 Conclusion
In this work, we first expose that audio-visual active speaker detection models are highly susceptible to adversarial attacks through comprehensive experiments, by investigating the white-box and black-box adversaries, single- and multi-modal attacks, training-aware, and inference-aware attack scenarios, and three attack algorithms with several attack budgets. Also, we propose the audio-visual interaction loss to enlarge the inter-class difference and intra-class similarity, resulting in more robust AVASD models for which budget-limited attackers can not find feasible adversarial samples. The experimental results illustrate that the proposed method is far more effective than adversarial training, and the proposed AVIL can complement adversarial training to further alleviate the adversarial vulnerability of AVASD models. In the future, we will investigate hyperparameter searching strategies to further improve the effectiveness of the proposed AVIL.
6 ACKNOWLEDGMENTS
This work was partially supported by the Ministry of Science and Technology, Taiwan (Grant no. MOST109-2221- E-002-163-MY3), and the Centre for Perceptual and Interactive Intelligence, an InnoCentre of The Chinese University of Hong Kong. This work was done while Haibin Wu was a visiting student at CUHK. Haibin Wu is supported by Google PHD Fellowship Scholarship.
References
- [1] Joseph Roth et al., “Ava-activespeaker: An audio-visual dataset for active speaker detection,” in ICASSP, 2020.
- [2] Joseph Roth et al., “Ava active speaker: An audio-visual dataset for active speaker detection,” in ICASSP 2020. IEEE, 2020, pp. 4492–4496.
- [3] Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus, “Intriguing properties of neural networks,” arXiv preprint arXiv:1312.6199, 2013.
- [4] Felix Kreuk, Yossi Adi, Moustapha Cisse, and Joseph Keshet, “Fooling end-to-end speaker verification with adversarial examples,” in ICASSP 2018. IEEE, 2018, pp. 1962–1966.
- [5] Haibin Wu, Xu Li, Andy T Liu, Zhiyong Wu, Helen Meng, and Hung-yi Lee, “Adversarial defense for automatic speaker verification by cascaded self-supervised learning models,” in ICASSP 2021. IEEE, 2021, pp. 6718–6722.
- [6] Arindam Jati, Chin-Cheng Hsu, Monisankha Pal, Raghuveer Peri, Wael AbdAlmageed, and Shrikanth Narayanan, “Adversarial attack and defense strategies for deep speaker recognition systems,” Computer Speech & Language, vol. 68, pp. 101199, 2021.
- [7] Haibin Wu, Yang Zhang, Zhiyong Wu, Dong Wang, and Hung-yi Lee, “Voting for the right answer: Adversarial defense for speaker verification,” arXiv preprint arXiv:2106.07868, 2021.
- [8] Yi Xie, Zhuohang Li, Cong Shi, Jian Liu, Yingying Chen, and Bo Yuan, “Real-time, robust and adaptive universal adversarial attacks against speaker recognition systems,” Journal of Signal Processing Systems, vol. 93, no. 10, pp. 1187–1200, 2021.
- [9] Yi Xie, Cong Shi, Zhuohang Li, Jian Liu, Yingying Chen, and Bo Yuan, “Real-time, universal, and robust adversarial attacks against speaker recognition systems,” in ICASSP 2020. IEEE, 2020, pp. 1738–1742.
- [10] Hadi Abdullah, Kevin Warren, Vincent Bindschaedler, Nicolas Papernot, and Patrick Traynor, “Sok: The faults in our asrs: An overview of attacks against automatic speech recognition and speaker identification systems,” in 2021 IEEE symposium on security and privacy (SP). IEEE, 2021, pp. 730–747.
- [11] Mirko Marras, Pawel Korus, Nasir D Memon, and Gianni Fenu, “Adversarial optimization for dictionary attacks on speaker verification.,” in Interspeech, 2019, pp. 2913–2917.
- [12] Haibin Wu, Xu Li, Andy T Liu, Zhiyong Wu, Helen Meng, and Hung-yi Lee, “Improving the adversarial robustness for speaker verification by self-supervised learning,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 30, pp. 202–217, 2021.
- [13] Rohan Kumar Das, Xiaohai Tian, Tomi Kinnunen, and Haizhou Li, “The attacker’s perspective on automatic speaker verification: An overview,” arXiv preprint arXiv:2004.08849, 2020.
- [14] Haibin Wu, Po-Chun Hsu, Ji Gao, Shanshan Zhang, Shen Huang, Jian Kang, Zhiyong Wu, Helen Meng, and Hung-Yi Lee, “Adversarial sample detection for speaker verification by neural vocoders,” in ICASSP 2022. IEEE, 2022, pp. 236–240.
- [15] Zhuohang Li, Cong Shi, Yi Xie, Jian Liu, Bo Yuan, and Yingying Chen, “Practical adversarial attacks against speaker recognition systems,” in Proceedings of the 21st international workshop on mobile computing systems and applications, 2020, pp. 9–14.
- [16] Zhiyuan Peng, Xu Li, and Tan Lee, “Pairing weak with strong: Twin models for defending against adversarial attack on speaker verification.,” in Interspeech, 2021, pp. 4284–4288.
- [17] Weiyi Zhang, Shuning Zhao, Le Liu, Jianmin Li, Xingliang Cheng, Thomas Fang Zheng, and Xiaolin Hu, “Attack on practical speaker verification system using universal adversarial perturbations,” in ICASSP 2021. IEEE, 2021, pp. 2575–2579.
- [18] Hao Tan, Le Wang, Huan Zhang, Junjian Zhang, Muhammad Shafiq, and Zhaoquan Gu, “Adversarial attack and defense strategies of speaker recognition systems: A survey,” Electronics, vol. 11, no. 14, pp. 2183, 2022.
- [19] Songxiang Liu et al., “Adversarial attacks on spoofing countermeasures of automatic speaker verification,” in ASRU 2019. IEEE, 2019, pp. 312–319.
- [20] Haibin Wu, Andy T Liu, and Hung-yi Lee, “Defense for black-box attacks on anti-spoofing models by self-supervised learning,” arXiv preprint arXiv:2006.03214, 2020.
- [21] Bowen Zhang, Benedetta Tondi, and Mauro Barni, “Adversarial examples for replay attacks against cnn-based face recognition with anti-spoofing capability,” Computer Vision and Image Understanding, vol. 197, pp. 102988, 2020.
- [22] Andre Kassis and Urs Hengartner, “Practical attacks on voice spoofing countermeasures,” arXiv preprint arXiv:2107.14642, 2021.
- [23] Haibin Wu, Songxiang Liu, Helen Meng, and Hung-yi Lee, “Defense against adversarial attacks on spoofing countermeasures of asv,” in ICASSP 2020. IEEE, 2020, pp. 6564–6568.
- [24] Nicholas Carlini and David Wagner, “Audio adversarial examples: Targeted attacks on speech-to-text,” in 2018 IEEE Security and Privacy Workshops (SPW). IEEE, 2018, pp. 1–7.
- [25] Hiromu Yakura and Jun Sakuma, “Robust audio adversarial example for a physical attack,” arXiv preprint arXiv:1810.11793, 2018.
- [26] Rohan Taori, Amog Kamsetty, Brenton Chu, and Nikita Vemuri, “Targeted adversarial examples for black box audio systems,” in 2019 IEEE security and privacy workshops (SPW). IEEE, 2019, pp. 15–20.
- [27] Yao Qin, Nicholas Carlini, Garrison Cottrell, Ian Goodfellow, and Colin Raffel, “Imperceptible, robust, and targeted adversarial examples for automatic speech recognition,” in International conference on machine learning. PMLR, 2019, pp. 5231–5240.
- [28] Moustafa Alzantot, Bharathan Balaji, and Mani Srivastava, “Did you hear that? adversarial examples against automatic speech recognition,” arXiv preprint arXiv:1801.00554, 2018.
- [29] Lea Schönherr, Katharina Kohls, Steffen Zeiler, Thorsten Holz, and Dorothea Kolossa, “Adversarial attacks against automatic speech recognition systems via psychoacoustic hiding,” arXiv preprint arXiv:1808.05665, 2018.
- [30] Chao-Han Yang, Jun Qi, Pin-Yu Chen, Xiaoli Ma, and Chin-Hui Lee, “Characterizing speech adversarial examples using self-attention u-net enhancement,” in ICASSP 2020. IEEE, 2020, pp. 3107–3111.
- [31] Juncheng B Li, Shuhui Qu, Xinjian Li, Po-Yao Bernie Huang, and Florian Metze, “On adversarial robustness of large-scale audio visual learning,” in ICASSP 2022. IEEE, 2022, pp. 231–235.
- [32] Shaojin Ding et al., “Personal vad: Speaker-conditioned voice activity detection,” arXiv preprint arXiv:1908.04284, 2019.
- [33] Abhishek Sehgal and Nasser Kehtarnavaz, “A convolutional neural network smartphone app for real-time voice activity detection,” IEEE Access, vol. 6, pp. 9017–9026, 2018.
- [34] Punarjay Chakravarty, Sayeh Mirzaei, Tinne Tuytelaars, and Hugo Van hamme, “Who’s speaking? audio-supervised classification of active speakers in video,” in Proceedings of the 2015 ACM on International Conference on Multimodal Interaction, 2015, pp. 87–90.
- [35] Muhammad Shahid, Cigdem Beyan, and Vittorio Murino, “S-vvad: visual voice activity detection by motion segmentation,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2021, pp. 2332–2341.
- [36] Ariel Ephrat, Inbar Mosseri, Oran Lang, Tali Dekel, Kevin Wilson, Avinatan Hassidim, William T Freeman, and Michael Rubinstein, “Looking to listen at the cocktail party: A speaker-independent audio-visual model for speech separation,” arXiv preprint arXiv:1804.03619, 2018.
- [37] Vicente Peruffo Minotto, Claudio Rosito Jung, and Bowon Lee, “Multimodal multi-channel on-line speaker diarization using sensor fusion through svm,” IEEE Transactions on Multimedia, vol. 17, no. 10, pp. 1694–1705, 2015.
- [38] Okan Köpüklü, Maja Taseska, and Gerhard Rigoll, “How to design a three-stage architecture for audio-visual active speaker detection in the wild,” in Proceedings of IEEE/CVF ICCV, 2021, pp. 1193–1203.
- [39] Sourya Roy, Kyle Min, Subarna Tripathi, Tanaya Guha, and Somdeb Majumdar, “Learning spatial-temporal graphs for active speaker detection,” arXiv preprint arXiv:2112.01479, 2021.
- [40] Yuanhang Zhang, Susan Liang, Shuang Yang, Xiao Liu, Zhongqin Wu, Shiguang Shan, and Xilin Chen, “Unicon: Unified context network for robust active speaker detection,” in Proceedings of the 29th ACM International Conference on Multimedia, 2021, pp. 3964–3972.
- [41] Baptiste Pouthier, Laurent Pilati, Leela K Gudupudi, Charles Bouveyron, and Frederic Precioso, “Active speaker detection as a multi-objective optimization with uncertainty-based multimodal fusion,” arXiv preprint arXiv:2106.03821, 2021.
- [42] Triantafyllos Afouras, Andrew Zisserman, et al., “Sub-word level lip reading with visual attention,” arXiv preprint arXiv:2110.07603, 2021.
- [43] Juan Léon Alcázar, Fabian Caba, Ali K Thabet, and Bernard Ghanem, “Maas: Multi-modal assignation for active speaker detection,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 265–274.
- [44] Joon Son Chung, “Naver at activitynet challenge 2019–task b active speaker detection (ava),” arXiv preprint arXiv:1906.10555, 2019.
- [45] Juan León Alcázar, Fabian Caba, Long Mai, Federico Perazzi, Joon-Young Lee, Pablo Arbeláez, and Bernard Ghanem, “Active speakers in context,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 12465–12474.
- [46] Juan León-Alcázar, Fabian Caba Heilbron, Ali Thabet, and Bernard Ghanem, “Maas: Multi-modal assignation for active speaker detection,” arXiv preprint arXiv:2101.03682, 2021.
- [47] Yuan-Hang Zhang, Jingyun Xiao, Shuang Yang, and Shiguang Shan, “Multi-task learning for audio-visual active speaker detection,” The ActivityNet Large-Scale Activity Recognition Challenge, pp. 1–4, 2019.
- [48] Ruijie Tao, Zexu Pan, Rohan Kumar Das, Xinyuan Qian, Mike Zheng Shou, and Haizhou Li, “Is someone speaking? exploring long-term temporal features for audio-visual active speaker detection,” in Proceedings of the 29th ACM International Conference on Multimedia, 2021, pp. 3927–3935.
- [49] Man-Wai Mak and Jen-Tzung Chien, Machine learning for speaker recognition, Cambridge University Press, 2020.
- [50] Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu, “Towards deep learning models resistant to adversarial attacks,” arXiv preprint arXiv:1706.06083, 2017.
- [51] A Kurakin, I Goodfellow, and S Bengio, “Adversarial examples in the physical world. arxiv 2016,” arXiv preprint arXiv:1607.02533, 2016.
- [52] Yinpeng Dong, Fangzhou Liao, Tianyu Pang, Hang Su, Jun Zhu, Xiaolin Hu, and Jianguo Li, “Boosting adversarial attacks with momentum,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 9185–9193.
- [53] Seyed-Mohsen Moosavi-Dezfooli, Alhussein Fawzi, and Pascal Frossard, “Deepfool: a simple and accurate method to fool deep neural networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 2574–2582.
- [54] Ian J Goodfellow, Jonathon Shlens, and Christian Szegedy, “Explaining and harnessing adversarial examples,” arXiv preprint arXiv:1412.6572, 2014.