1. Introduction
Recently, machine learning models are widely used in various fields, but many security issues also exist, especially, the adversarial attack. Studies show that several machine learning models are vulnerable to adversarial examples which generated by adding an elaborately designed perturbations to a benign sample [1,2].
For automatic speech recognition (ASR) system, its accuracy can be improved by deep learning network. However, potential security risk is also aroused. A benign audio could be decoded as a harmful command by adversarial attack. Existing defense methods against audio adversarial examples have following deficiencies: necessity to retrain ASR models [3] or use additional machine learning models [4-6], only applicable to specific models [7,8]. Most of those methods may be inspired by image adversarial example domain and not consider the characteristics of ASR systems.
The main contributions of our work: 1) Utilizing the characteristics of ASR systems, we offer a novel insight in explore countermeasures against audio adversarial example. 2) We propose a simple, generic and efficient method against on audio adversarial attack. Our method does not need to retrain ASR systems, and can be applied for different ASR systems (such as Classification, Kaldi and DeepSpeech [9-11]) and achieves better performance than existing methods. 3) For different scenarios, we give a variety of application strategies of our method.
The rest of the paper is organized as follows. In Section II, related works of the adversarial attacks and countermeasures for ASR system are described. Details of the proposed method, which includes the frame offsets, defense strategies and theoretical analysis, are described in Section III. The evaluation results of our method are present in Section IV. We conclude our work in Section V.
2. Related Work
According to the output types of different ASR systems, the attack task can be divided into two categories: speech-to-label and speech-to-text. For speech-to-label task, an audio is discriminated to a label by ASR classification model. The adversary is targeted at making the audio be discriminated to a label which is different from the real one. Alzantot et al. [12] proposed a genetic algorithm-based method to generate audio adversarial examples against speech-to-label model. For speech-to-text task, an audio is directly decoded as a text by ASR system. The adversary is targeted at making the audio be decoded as a pre-specified text. Carlini et al. [13] are the first to make the audio adversarial examples work on speech-to-text models. Yuan et al. [14] choose the music as the carrier and achieved practical over-the-air audio adversarial attacks. Several of the later works are based on [13] to improve the generation efficiency and robust of adversarial example [15-17].
On the other hand, there are several countermeasures to audio adversarial example. Sun et al. [3] add adversarial examples to the training dataset to make the ASR system more robust. Samizade et al. [6] design a CNN based classification model to detect audio adversarial example. Some other methods denoise the audio adversarial example with self-attention U- Net [5] or GAN [4] to fail the adversarial attacks. Yang et al. [18] propose a novel detection method using temporal dependency. Several transformation-based methods [18,19] (i.e. down-sampling, quantization, local-smoothing, compressions et al.) are utilized to defense the audio adversarial examples.
3. Proposed Method
The frame offsets [20] are defined as the shifting samples of the frame grid between the first and second encoding. In our works, we make an audio get frame offsets by appending silence clip (ASC) at the beginning of it. In the section, we first introduce how frame offsets resists the audio adversarial examples, then give three application strategies for different scenarios and finally give a theoretical analysis to which length silence clip is appropriate for appending.
3.1 Frame offsets with ASC
Most of modern ASR systems often need to extract features (e.g., MFCCs) from audio and the value of the features will largely determine the output of the system. It should be noted that the extracted features will be affected by the window size and frame stride. As shown in Fig. 1, if the first frame gets offsets by appending a silence clip at the beginning of an audio, the remaining frames will also get the same offsets and the values of extracted features will be changed. For a benign audio, the recognition results will not be affected by frame offsets. Generally, at the beginning of recording, there will be a silence clip with a random length. The length of the silence clip will not affect the recognition output due to the temporal dependency [18] which has been widely used in audio systems. For adversarial audio, the adversarial perturbation is elaborately designed for the original audio temporal. When the temporal is changed by frame offsets, the perturbations will degenerate into ordinary noise. Fig. 2 shows the change in recognition results of benign audio and adversarial audio after frame offsets with ASC.
Fig. 1. Comparison of framing process for an audio (left) and its processed sample (right) by frame offsets with ASC
Fig. 2. Recognition results of benign audio and adversarial audio after frame offsets with ASC Fig. 3. Pipeline of proposed method with different strategies
3.2 Exploit frame offsets by different strategies
As mentioned in Section 3.1, the adversarial example is more vulnerable to frame offsets. The phenomenon can be exploited by different strategies such as defending, detecting and hybrid strategy as the countermeasure against audio adversarial example.
3.2.1 Defending
The purpose of defending strategy is to make the adversarial audio harmless. The first row of Fig. 3 shows that the pipeline of defending strategies. Given an audio instance 𝑥 (benign sample or adversarial sample), we append an appropriate length silence clip at the beginning of the audio and fed it into the ASR system. For benign audio, the recognition results will not be changed. For adversarial audio, the recognition results will different from before and harmless. We will explore which length silence clip is appropriate for appending by theoretical analysis and experiments.
Fig. 3. Pipeline of proposed method with different strategies
3.2.2 Detecting
The purpose of detecting strategy is to identify whether unknow types of audio are adversarial samples. The recognition result of adversarial example is more likely to be affected by frame offsets. Hence, we can use the amount of change in recognition results before and after using frame offsets to detect whether an audio sample is adversarial sample. Given an audio instance 𝑥 (benign sample or adversarial sample), ASR function 𝑔(⋅) and manipulation function 𝐴𝑆𝐶𝜖(⋅). 𝐴𝑆𝐶𝜖(𝑥) means to append a silence clip at the beginning of 𝑥. The details of 𝐴𝑆𝐶𝜖(⋅ ) will be discussed in Section 3.2.3. We use the word/character change rate (𝐶𝑅) to measure the amount of change in recognition results of 𝑥, 𝐶𝑅 is defined as follow:
\(C R=\frac{\min \left(D\left(g\left(A S C_{\epsilon}(x)\right), g(x)\right), L\right)}{L}\) (1)
where 𝐿 denotes the number of words/characters of 𝑔(𝑥), 𝐷(⋅, ⋅) denote the distance of two texts ((𝐴𝑆𝐶𝜖(𝑥)) and 𝑔(𝑥) in our case). The word error rate (𝑊𝐸𝑅) and character error rate (𝐶𝐸𝑅) [21] is used as the distance function 𝐷(⋅, ⋅). Corresponding, we can get the word change rate (𝑊𝐶𝑅) and character change rate (𝐶𝐶𝑅). The magnitude of 𝐶𝑅 can characterize the possibility whether 𝑥 is an adversarial example. The closer 𝐶𝑅 is to 1, the more likely 𝑥 is (1) regarded as adversarial example.
3.2.3 Hybrid strategy
Different strategies may be suitable for different scenarios. The defending strategy is straightforward, but the recognition result of benign audio also be modified. The detecting strategy does not modify the recognition result, but the users may have a bad experience when the adversarial attack occurs during transmission and the users are not aware of it. Hence, we can combine the defending strategy and detecting strategy. The last row of Fig. 3 shows the pipeline of hybrid strategy. Firstly, it needs to detect whether an audio is adversarial example. If the audio is an adversarial example, we can use frame offsets or other defending methods to defend it.
3.3 Theoretical Analysis
Given an audio signal 𝑥, with 𝑓 Hz sampling rate, 𝑡 seconds and 𝑙 samples (𝑙=𝑓∗𝑡). When extracting features, different audio system could use different window size and stride. We suppose that the window size is 𝑡𝑤 seconds and the stride is 𝑡𝑠 second. One window has 𝑙𝑤= 𝑓∗𝑡𝑤 samples and one stride has 𝑙𝑠=𝑓∗𝑡𝑠 samples. Splitting 𝑥 into a series of frames:
\(x=\left[x^{(1)}, \cdots, x^{(N)}\right], \quad N=\left\lfloor\frac{l-\left(l_{w}-l_{s}\right)}{l_{s}}\right\rfloor\) (2)
\(\boldsymbol{x}^{(k)}=\left[\boldsymbol{x}_{i+1}, \cdots, \boldsymbol{x}_{i+l_{w}}\right], \quad i=(k-1) * l_{s}\) (3)
where 𝒙 is vector-notation of 𝑥, 𝒙(𝑘) denotes the 𝑘-th frame of 𝑥 and 𝒙𝑖 denotes the 𝑖-th sample of 𝑥. Finally, we can get N frames.
In order to offset the frames of 𝑥, we append a silence clip at the beginning of 𝑥. Let us formula the operation as follows:
\(\operatorname{ASC}_{\epsilon}(x)=[\underbrace{0, \cdots, 0}_{l_{a}}, x], \quad l_{a}=\left\lfloor\epsilon * l_{s}\right\rfloor\) (4)
where 𝐴𝑆𝐶𝜖(⋅) is the frame offset function and 𝜖 is the coefficient which is used to control the length of silence clip (𝑙𝑎).
We can get 𝒙̂ by feeding𝒙 into 𝐴𝑆𝐶𝜖(⋅):
\(\widehat{x}=A S C_{\epsilon}(x)\) (5)
\(\widehat{\boldsymbol{x}}=\left[\widehat{\boldsymbol{x}}^{(1)}, \cdots, \widehat{\boldsymbol{x}}^{(\bar{N})}\right], \quad \hat{N}=\left\lfloor\frac{l+l_{a}-\left(l_{w}-l_{s}\right)}{l_{s}}\right]\) (6)
\(\widehat{\boldsymbol{x}}^{(k)}=\left[\hat{\boldsymbol{x}}_{i+1}, \cdots, \widehat{\boldsymbol{x}}_{i+l_{w}}\right], \quad i=(k-1) * l_{s}\) (7)
Via 𝒙̂𝑖=𝒙𝑖−𝑙𝑎, we can get:
\(\hat{x}^{(k)}= \begin{cases}\underbrace{[0, \cdots, 0]}_{l_{w}}], & k \leq\left\lfloor\frac{l_{a}}{l_{s}}\right\rfloor \\ {[\underbrace{0, \cdots, 0}_{l_{a} m o d l_{s}}, x_{1}, \cdots, x_{l_{w}-l_{a}}],} & k \leq\left\lfloor\frac{L_{a}}{l_{s}}\right\rfloor+1 \\ {[\underbrace{x_{i-l_{a}+1}, \cdots, x_{i}}_{l_{a} \bmod l_{s}}, x_{i+1}, \cdots, x_{i+l_{w}-l_{a}}],} & k \leq\left\lfloor\frac{L_{a}}{l_{s}}\right]+1\end{cases}\) (8)
For any 𝒙(𝑘), we can find \(\widehat{\boldsymbol{x}}^{(k+}\left[\frac{l_{a}}{l_{s}} \mid\right)_{\text {or }} \widehat{\boldsymbol{x}}^{\left(k+\left|\frac{l_{a}}{l_{s}}\right|+1\right)}\) that is closest to 𝒙(𝑘) in 𝒙̂. We denote 𝐿𝐷𝑃(⋅, ⋅) as the length of different part for two frames. Then we can get:
\(L D P_{1}=L D P\left(\boldsymbol{x}^{(k)}, \widehat{\boldsymbol{x}}^{\left(k+\left|\frac{l_{a}}{l_{s}}\right|\right)}\right)=l_{a} \bmod l_{s}\) (9)
\(L D P_{2}=L D P\left(\boldsymbol{x}^{(k)}, \widehat{\boldsymbol{x}}^{\left(k+\left[\frac{l_{a}}{l_{s}} \mid+1\right)\right.}\right)=l_{s}-\left(l_{a} \bmod l_{s}\right)\) (10)
\(L D P_{\min }=\min \left(L D P_{1}, L D P_{2}\right)\) (11)
where 𝐿𝐷𝑃min is the minimal length of different part for 𝒙(𝑘) and is the frame which is closest to 𝒙(𝑘) in 𝒙̂.
𝐿𝐷𝑃𝑚𝑖𝑛 can measure the degree to which the value of features changes after frame offsets with different 𝜖. Fig. 4 shows that the value of 𝐿𝐷𝑃𝑚𝑖𝑛 via different frame offsets. It presents a certain periodicity, which means appending more silence frame may make no sense. In this work, the range of 𝜖 is set to [0,1]. When \(\epsilon=\frac{1}{2}\), our method could obtain a best performance.
Fig. 4. 𝐿𝐷𝑃min via different frame offsets \(\left(l_{a}=\left\lfloor\epsilon * l_{s}\right\rfloor\right)\)
4. Experimental Results
4.1 Experimental Setup
4.1.1 Attack Method
For speech-to-label task, we evaluate on the genetic algorithm-based attack (GA). The Commander Song attack (CommanderSong) and the optimization-based attack (OPT) are evaluated for speech-to-text task. In our experiment, the audio adversarial samples which are generated by these attack methods are sent directly to the ASR system.
Different attack methods use different ASR models as the threat model respectively. We used For GA, a convolutional speech commands classification model is used as same in [12]. For CommanderSong attack, we evaluate the performance on Kaldi speech recognition platform. For OPT attack, we use DeepSpeech which is a biRNN based speech-to-text transcription network.
GA: GA is a state-of-the-art speech-to-label attack proposed in [12]. Here an audio classification model is attacked and the output consists of 10 different labels. They aimed to attack such a network to misclassify an adversarial audio based on either targeted or untargeted attack.
CommanderSong: CommanderSong [14] is a speech-to-text targeted attack which use songs as the original audio. The adversarial audio can even be played over the air with its adversarial characteristics. Since the source codes of CommanderSong are not available, we evaluate on the generated adversarial audios provided by the authors.
OPT: We consider the targeted speech-to-text attack proposed by [13], which uses CTC- loss in a speech recognition system as an objective function and solves the task of adversarial attack as an optimization problem.
4.1.2 Dataset
Speech Commands: Speech Commands dataset contains 65000 audio files. Each audio is a single command and has one second duration. In this work, we choose 10 types commands. the commands are “yes”, “no”, “up”, “ own”, “left”, “right”, “on”, “off”, “stop”, and “go”.
LibriSpeech: LibriSpeech is a corpus of approximately 1000 hours of 16Khz English speech. The data is derived from read audiobooks from the LibriVox project. In this work, we use its test-clean dataset in their website [22].
Common Voice: Common Voice is a free audio dataset provided by Mozilla for ASR system. This dataset is public and contains samples from human speaking audio files. In this work, we used its subset which is 16Khz-sampled and has 3.998s average duration. The dataset can be found in [23].
Timit: Timit dataset contains 6300 audio files and consists of only 10 sentences. Each sentence is 30 seconds long and is spoken by 630 different speakers. In this work, wo use its first 100 sample to generate adversarial audio.
4.1.3 Compared Method
For defense method, Down Sampling, Local Smoothing and Quantization are considered here. For detection method, a novel method using temporal dependency method (TD Method) is considered.
Down Sampling: Based on sampling theory, it is possible to down-sample a band-limited audio file without sacrificing the quality of the recovered signal while mitigating the adversarial perturbations in the reconstruction phase. In this work, we first down-sample the original 16kHz audio to 8kHz, then up-sample the audio to 16kHz again.
Local Smoothing: We use a sliding window with a fixed length for local smoothing to reduce the adversarial perturbation. Given an audio sample 𝑥, we consider the 2𝐾+1samples window which is denoted by [𝒙𝑖−𝐾+1, ⋯, 𝒙𝑖, ⋯, 𝒙𝑖+𝐾−1], and replace 𝑥𝑖 by the smoothed value (median in our case) of the window.
Quantization: Since the amplitude of adversarial perturbation is usually small in the input space, it could be disrupted by rounding the amplitude of audio sampled data into the nearest integer multiple of 𝑞. In this work, we choose 𝑞 = 256, 512 which obtains the best performance in [18].
TD Method: TD Method [18] is a detection method which can exploits the temporal dependency property of audio data to detect audio adversarial examples.
4.1.4 Evaluation Metrics
𝑨𝑺𝑹𝒂𝒗𝒈: 𝐴𝑆𝑅𝑎𝑣𝑔 is an average value of attack success rate for every type GA attack. After applying a defense method to input samples, 𝐴𝑆𝑅𝑎𝑣𝑔 will decrease. A low 𝐴𝑆𝑅𝑎𝑣𝑔 means that the defense method has good performance.
𝑹𝒃𝒆𝒏𝒊𝒈𝒏/𝑹𝒂𝒅𝒗: To evaluate the effectiveness of transformation methods and our method against speech-to-text attack, we report the ratio of translation distance between instance and corresponding ground truth before and after transformation. 𝑅𝑏𝑒𝑛𝑖𝑔𝑛is the effectiveness ratio for benign instances. 𝑅𝑎𝑑𝑣 is the similar effectiveness ratio for adversarial audio.
\(R_{\text {benign }}=\frac{D\left(g\left(T\left(x_{\text {benign }}\right)\right), y\right)}{D\left(g\left(x_{\text {benign }}\right), y\right)}, \quad R_{a d v}=\frac{D\left(g\left(T\left(x_{a d v}\right)\right), y\right)}{D\left(g\left(x_{a d v}\right), y\right)}\) (12)
where 𝑥𝑏𝑒𝑛𝑖𝑔𝑛 denotes a benign audio, 𝑥𝑎𝑑𝑣 denotes an adversarial audio, 𝑦 is a text of the ground truth, D(⋅, ⋅) denotes the distance function (WER and CER in our case) and 𝑇(⋅) denotes the input transformation function (i.e. down-sampling, quantization, local-smoothing, compressions et al.). Particularly, in our method, 𝐴𝑆𝐶𝜖(⋅) is used as 𝑇(⋅).
AUC score: For detection task, AUC score is a commonly used metric. It stands for the area under the ROC Curve. In this work, we use CR which is mentioned in Section 3.2.2 as the output probability to calculate AUC score.
4.2 Length of Silence Clip
In this section, we explore which length silence clip is appropriate to appending to achieve a better defending performance. We first evaluate on DeepSpeech by taking a benign sample and using OPT attack to generate an adversarial sample, then keep appending silence clip at the beginning of the both audios and record the value of CER. Fig. 5 shows the results of this process. Both of these curves show a certain periodicity and have the same trend with Fig. 4. The CER fluctuation range of the benign sample is smaller than the adversarial sample, which indicates that the adversarial examples is more vulnerable to frame offsets.
Fig. 5. CER via different frame offsets (𝑙𝑎=⌊𝜖∗𝑙𝑠⌋): (a) benign sample (b) adversarial sample
4.3 Defense Result
In this section, we will measure the performance of our method of defense strategy on three different attacks: GA, CommanderSong and OPT. For comparison, we also measure the performance of input transformation methods which are proposed in [18,19].
GA: We first evaluate our method on the GA attack [12] which is a speech-to-label type attack. We choose 10 types samples from SpeechCommand dataset. For each type we use the remaining other types as the target to generate adversarial example, every type attack has 50 samples. The average of attack success rate (𝐴𝑆𝑅𝑎𝑣𝑔) is 83% without defense. From the source code of the Classification model, we can know the frame stride is 10ms, so we append 5ms \(\left(\epsilon=\frac{1}{2}\right)\) silence clip at the beginning of samples. Finally, by using our method, the 𝐴𝑆𝑅𝑎𝑣𝑔 fall to 5.6% and is lower than input transformation method, the result of all method measured here is listed in Table 1, and the detailed results for every type attack of our method shown in Fig. 6.
Table 1. Evaluation results for defense with different on Classification ASC
Fig. 6. Attack success rate (a) without defense for every type attack, (b) with our defense method.
CommanderSong: We also evaluate our method on CommanderSong attack [12]. The dataset we used here can be found in [24]. It contains only 10 samples, 5 samples of which are generated by WTA attack and the other samples are generated by WAA attacks. Due to too few samples, 𝐴𝑆𝑅𝑎𝑣𝑔 cannot be calculated accurately. In Table 2, We list the recognition results of some adversarial examples in this dataset before and after using our method. As shown in Table 2, our method can effectively defend the ConmmanderSong attack. Table 2. Evaluation results for defense on CommanderSong attack.
Table 2. Evaluation results for defense on CommanderSong
OPT: Eventually, we evaluate our defense method on OPT attack (OPT) [13] which is a text-to-speech type attack, and DeepSpeech is used as the victim model. We choose the effectiveness ration for benign instances (𝑅𝑏𝑒𝑛𝑖𝑔𝑛) and for adversarial sample (𝑅𝑎𝑑𝑣) as the metrics, and experiments on three different datasets (LibriSpeech, CommonVoice and Timit). The results of our method and input transformation method are listed in Table 3. The 𝑅𝑏𝑒𝑛𝑖𝑔𝑛 of our method is smaller than other methods and closer to 1.0. This means that the manipulation of our method to audio has little effect on the recognition accuracy of benign instances. Table 4 gives some other results when 𝜖 takes other values. When \(\epsilon=\frac{1}{2}\), our method achieved the best result, and this coincides with our conjecture in Chapter 3. With 𝜖 gets bigger, the experimental results start to deteriorate. Particularly, when 𝜖 up to 1, we get the worst result. Because this is only equivalent to adding 1 frame in front of samples and does not offset the remaining frames. 𝑅𝑎𝑑𝑣 is close to 1 and the recognition result of adversarial example dose not change much. Considering such a countermeasure against our method: if 𝜖 is fixed, the attacker can add 1−𝜖 silence clip to the generated adversarial example and the true 𝜖will become 1. Hence, we randomly selected 𝜖 in the range from \(\frac{1}{4} \text { to } \frac{3}{4}\), the results show that our method is still effective on this condition.
Table 3. Evaluation results for defense with different on DeepSpeech
Table 4. Evaluation results for defense with different 𝜖 on DeepSpeech
4.4 Detection Result
In this section, we will measure the performance of our method for detection on OPT. Because the output of the Classification model is only a label that can't be calculated the distance by 𝐷(⋅, ⋅), so GA is not considered here. For comparison, we also measure the TD method that is proposed by [18] in the same experimental setting. Different from defense, the metrics used for detection are AUC score. Table 5 summarizes the evaluation results of detection strategy.
Table 5. Evaluation results for detection with different method on DeepSpeech
OPT: Here we use AUC score (with WER and CER) as the metric, and evaluate on three different datasets. All the target texts we used is "This is an adversarial example". Table 5 shows that the performance of our method is better than TD Method and the AUC scores in our method are close to 1 no matter on what datasets. Table 6 gives some other results when 𝜖 takes other values. When \(\epsilon=\frac{1}{3}, \frac{1}{2}, \frac{3}{4}\) the AUC score of our method is close to 1. Particularly, when 𝜖 up to 1, we get the worst result.
Table 6. Evaluation results for detection with different 𝜖 on DeepSpeech
5. Conclusion
In this work, we proposed a simple, generic and efficient method against audio adversarial attack. For different scenarios, we give a variety of usage strategies (defend, detect and hybrid strategy) of our method. Evaluating on three state-of-the-arts adversarial attacks against on different ASR systems respectively, the results demonstrate that the proposed method can effective improving the robust of audio systems.
Although our method can detect and defend the audio adversarial example, the perturbations still exist as noise which will affect the recognition accuracy of ASR system. The future work should focus on recovering the ground truth of original audio from adversarial audio. One possible way is to combine our method with some denoising methods. First, by using our method, the perturbations will degenerate into ordinary noise. Then, denoising methods can be used to reduce the noise.
For attack methods in this work, the audio samples are sent directly to the ASR system without playing in physical world. With the deepening of research in this field, future attack methods will be able to generate audio adversarial examples that are effective in physical world. Hence, the physical scene should also be considered in corresponding countermeasures.
References
- I. J. Goodfellow, J. Shlens, and C. Szegedy, "Explaining and harnessing adversarial examples," in Proc. of the 3rd International Confernce at Learn. Represent(ICLR), 2015.
- N. Carlini and D. Wagner, "Towards evaluating the robustness of neural networks," in Proc. of IEEE Symposium on Security and Privacy(SP), pp. 39-57, 2017.
- S. Sun, C. F. Yeh, M. Ostendorf, M. Y. Hwang, and L. Xie, "Training augmentation with adversarial examples for robust speech recognition," in Proc. of Annuual Conference International Speech Communncation Association(INTERSPEECH), pp. 2404-2408, 2018.
- S. Latif, R. Rana, and J. Qadir, "Adversarial machine learning and speech emotion recognition: utilizing generative adversarial networks for robustness," arXiv Prepr. arXiv1811.11402, 2018.
- C. Yang, J. Qi, P. Chen, X. Ma, and C. Lee, "Characterizing speech adversarial examples using self-attention U-Net enhancement," in Proc. of IEEE International Conference on Acoustics, Speech and Signal Processing(ICASSP), pp. 3107-3111, 2020.
- S. Samizade, Z. Tan, C. Shen and X. Guan, "Adversarial example detection by classification for deep speech recognition," in Proc. of IEEE International Conference on Acoustics, Speech and Signal Processing(ICASSP), pp. 3102-3106, 2020.
- K. Rajaratnam, K. Shah, and J. Kalita, "Isolated and ensemble audio preprocessing methods for detecting adversarial examples against automatic speech recognition," in Proc. of the 30th Conference on Computational Linguistics and Speech Processing(ROCLING), pp. 16-30, 2018.
- P. Ma, S. Petridis, and M. Pantic, "Detecting adversarial attacks on audio-visual speech recognition," arXiv Prepr. arXiv1912.08639, 2019.
- T. N. Sainath and C. Parada, "Convolutional neural networks for small-footprint keyword spotting," in Proc. of INTERSPEECH 2015, pp. 1478-1482, 2015.
- Kaldi. [Online]. Available: https://kaldi-asr.org
- DeepSpeech. [Online]. Available: https://github.com/mozilla/DeepSpeech
- M. Alzantot, B. Balaji, and M. Srivastava, "Did you hear that? Adversarial examples against automatic speech recognition," arXiv Prepr. arXiv1801.00554, 2018.
- N. Carlini and D. Wagner, "Audio adversarial examples: Targeted attacks on speech-to-text," in Proc. of 2018 IEEE Symposium on Security and Privacy Workshops, pp. 1-7, 2018.
- X. Yuan, Y. Chen, Y. Long, X. Liu, K. Chen, H. Huang, and X. Wang, "CommanderSong: A systematic approach for practical adversarial voice recognition," in Proc. of the 27th USENIX Security Symposium, pp. 49-64, 2018.
- X. Liu, K. Wan, Y. Ding, X. Zhang, and Q. Zhu, "Weighted-sampling audio adversarial example attack," AAAI Technical Track: Machine Learning, vol. 34, no. 04, pp. 4908-4915, Apr. 2020.
- Y. Qin, N. Carlini, I. Goodfellow, G. Cottrell, and C. Raffel, "Imperceptible, Robust, and targeted adversarial examples for automatic speech recognition," in Proc. of the 36th International Conference on Machine Learning(ICML), pp. 9141-9150, 2019.
- L. Schonherr, K. Kohls, S. Zeiler, T. Holz, and D. Kolossa, "Adversarial attacks against automatic speech recognition systems via psychoacoustic hiding," in Proc. of Network and Distributed Systems Security(NDSS) Symposium, 2019.
- Z. Yang, P. Y. Chen, B. Li, and D. Song, "Characterizing audio adversarial examples using temporal dependency," in Proc. of the 7th International Conference on Learning Represent(ICLR) 2019.
- H. Kwon, H. Yoon, and K. W. Park, "Poster: Detecting audio adversarial example through audio modification," in Proc. of the ACM Conference on Computer and Communications Security, pp. 2521-2523, 2019.
- R. Yang, Z. Qu, and J. Huang, "Detecting digital audio forgeries by checking frame offsets," in Proc. of the 10th ACM Workshop on Multimedia and Security, 2008.
- V. I. Levenshtein, "Binary codes capable of correcting deletions, insertions, and reversals," Soviet Physics Doklady, vol. 10, no. 8, pp. 707-710, 1966.
- Open SLR. [Online]. Available: http://www.openslr.org/12
- Audio Adversarial Examples. [Online]. Available: https://nicholas.carlini.com/code/audio adversarial_examples
- Kaichen. [Online]. Available: http://kaichen.org