
Research Article
Towards Defending Adversarial Attacks with Temperature Regularization in Automatic Modulation Recognition
@INPROCEEDINGS{10.1007/978-3-031-23902-1_27, author={Tao Jiang and Huabao Xu and Linlin Liang and Peihan Qi}, title={Towards Defending Adversarial Attacks with Temperature Regularization in Automatic Modulation Recognition}, proceedings={Mobile Multimedia Communications. 15th EAI International Conference, MobiMedia 2022, Virtual Event, July 22-24, 2022, Proceedings}, proceedings_a={MOBIMEDIA}, year={2023}, month={2}, keywords={Deep learning Automatic modulation recognition Adversarial defense Temperature regularization}, doi={10.1007/978-3-031-23902-1_27} }
- Tao Jiang
Huabao Xu
Linlin Liang
Peihan Qi
Year: 2023
Towards Defending Adversarial Attacks with Temperature Regularization in Automatic Modulation Recognition
MOBIMEDIA
Springer
DOI: 10.1007/978-3-031-23902-1_27
Abstract
Deep learning has been shown to perform extremely well at various machine learning tasks. However, these same architectures are highly vulnerable to adversarial examples: malicious inputs carefully crafted by adversaries which can force a neural network to produce erroneous predictions with high confidence. This undermines the security of deep learning algorithms when apply to those security-sensitive applications. Existing works have shown that the Signal Modulation Recognition (SMR) solutions based on deep learning are also susceptible to adversarial attacks. In this paper, we propose a new approach called temperature regularization to defense a deep learning scheme against white-box attacks in signal modulation recognition. Specifically, we introduce different temperatures to the softmax layer during the training of the neural network. Experimental results show that training a neural network with an appropriate high temperature can significantly enhance its robustness to three white-box attacks.