CrossRef Open Access 2025

Testing the limits: exploring adversarial techniques in AI models

Apostolis Zarras Athanasia Kollarou Aristeidis Farao Panagiotis Bountakas Christos Xenakis

Abstrak

The rising adoption of artificial intelligence and machine learning in critical sectors underscores the pressing need for robust systems capable of withstanding adversarial threats. While deep learning architectures have revolutionized tasks such as image recognition, their susceptibility to adversarial techniques remains an open challenge. This article evaluates the impact of various adversarial methods, including the fast gradient sign method, projected gradient descent, DeepFool, and Carlini & Wagner, on five neural network models: a fully connected neural network, LeNet, Simple convolutional neural network (CNN), MobileNetV2, and VGG11. Using the E V AI SION tool explicitly developed for this research, these attacks were implemented and analyzed based on accuracy, F1-score, and misclassification rate. The results revealed varying levels of vulnerability across the tested models, with simpler architectures occasionally outperforming more complex ones. These findings emphasize the importance of selecting the most appropriate adversarial technique for a given architecture and customizing the associated attack parameters to achieve optimal results in each scenario.

Penulis (5)

A

Apostolis Zarras

A

Athanasia Kollarou

A

Aristeidis Farao

P

Panagiotis Bountakas

C

Christos Xenakis

Format Sitasi

Zarras, A., Kollarou, A., Farao, A., Bountakas, P., Xenakis, C. (2025). Testing the limits: exploring adversarial techniques in AI models. https://doi.org/10.7717/peerj-cs.3330

Akses Cepat

Lihat di Sumber doi.org/10.7717/peerj-cs.3330
Informasi Jurnal
Tahun Terbit
2025
Bahasa
en
Sumber Database
CrossRef
DOI
10.7717/peerj-cs.3330
Akses
Open Access ✓