A Comprehensive Review on Adversarial Attack Detection Analysis in Deep Learning
DOI:
https://doi.org/10.32628/CSEIT2361054Keywords:
Adversarial Attacks, Deep Learning, Detection Techniques, Machine Learning, Interpretability, Explainable Artificial Intelligence (XAI), Reinforcement Learning.Abstract
This comprehensive review investigates the escalating concern of adversarial attacks on deep learning models, offering an extensive analysis of state-of-the-art detection techniques. Encompassing traditional machine learning methods and contemporary deep learning approaches, the review categorizes and evaluates various detection mechanisms while addressing challenges such as the need for benchmark datasets and interpretability. Emphasizing the crucial role of explaining ability and trustworthiness, the paper also explores emerging trends, including the integration of technologies like explainable artificial intelligence (XAI) and reinforcement learning. By synthesizing existing knowledge and outlining future research directions, this review serves as a valuable resource for researchers, practitioners, and stakeholders seeking a nuanced understanding of adversarial attack detection in deep learning.
References
- G. Ryu and D. Choi, “Detection of adversarial attacks based on differences in image entropy,” International Journal of Information Security, 2023, doi: 10.1007/s10207-023-00735-6.
- X. Cui, “Targeting Image-Classification Model,” pp. 1–13, 2023.
- M. Kim and J. Yun, “AEGuard: Image Feature-Based Independent Adversarial Example Detection Model,” Security and Communication Networks, vol. 2022, 2022, doi: 10.1155/2022/3440123.
- P. Lorenz, M. Keuper, and J. Keuper, “Unfolding Local Growth Rate Estimates for (Almost) Perfect Adversarial Detection,” pp. 27–38, 2023, doi: 10.5220/0011586500003417.
- L. Shi, T. Liao, and J. He, “Defending Adversarial Attacks against DNN Image Classification Models by a Noise-Fusion Method,” Electronics (Switzerland), vol. 11, no. 12, 2022, doi: 10.3390/electronics11121814.
- A. S. Almuflih, D. Vyas, V. V Kapdia, M. R. N. M. Qureshi, K. M. R. Qureshi, and E. A. Makkawi, “Novel exploit feature-map-based detection of adversarial attacks,” Applied Sciences, vol. 12, no. 10, p. 5161, 2022.
- M. Khan et al., “Alpha Fusion Adversarial Attack Analysis Using Deep Learning,” Computer Systems Science and Engineering, vol. 46, no. 1, pp. 461–473, 2023, doi: 10.32604/csse.2023.029642.
- N. Ghaffari Laleh et al., “Adversarial attacks and adversarial robustness in computational pathology,” Nature Communications, vol. 13, no. 1, pp. 1–10, 2022, doi: 10.1038/s41467-022-33266-0.
- Y. Wang et al., “Adversarial Attacks and Defenses in Machine Learning-Powered Networks: A Contemporary Survey,” pp. 1–46, 2023, [Online]. Available: http://arxiv.org/abs/2303.06302
- H. Hirano, A. Minagi, and K. Takemoto, “Universal adversarial attacks on deep neural networks for medical image classification,” BMC Medical Imaging, vol. 21, no. 1, pp. 1–13, 2021, doi: 10.1186/s12880-020-00530-y.
- A. Talk, F. Wikipedia, A. Wikipedia, and C. Wikipedia, “University of Science and Technology of China,” no. 6, p. 29201, 2001.
- Y. Zheng and S. Velipasalar, “Part-Based Feature Squeezing To Detect Adversarial Examples in Person Re-Identification Networks,” Proceedings - International Conference on Image Processing, ICIP, vol. 2021-September, pp. 844–848, 2021, doi: 10.1109/ICIP42928.2021.9506511.
- B. Liang, H. Li, M. Su, X. Li, W. Shi, and X. Wang, “Detecting Adversarial Image Examples in Deep Neural Networks with Adaptive Noise Reduction,” IEEE Transactions on Dependable and Secure Computing, vol. 18, no. 1, pp. 72–85, 2021, doi: 10.1109/TDSC.2018.2874243.
- M. A. Ahmadi, R. Dianat, and H. Amirkhani, “An adversarial attack detection method in deep neural networks based on re-attacking approach,” pp. 10985–11014, 2021.
- K. Ren, T. Zheng, Z. Qin, and X. Liu, “Adversarial Attacks and Defenses in Deep Learning,” Engineering, vol. 6, no. 3, pp. 346–360, 2020, doi: 10.1016/j.eng.2019.12.012.
Downloads
Published
Issue
Section
License
Copyright (c) IJSRCSEIT

This work is licensed under a Creative Commons Attribution 4.0 International License.