The IRMA Community
Newsletters
Research IRM
Click a keyword to search titles using our InfoSci-OnDemand powered search:
|
From Attack to Defense: Strengthening DNN Text Classification Against Adversarial Examples
Abstract
In recent academic discussions surrounding the textual domain, there has been significant attention directed towards adversarial examples. Despite this focus, the area of detecting such adversarial examples remains notably under-investigated. In this chapter, the authors put forward an innovative approach for the detection of adversarial examples within the realm of natural language processing (NLP). This approach draws inspiration from the local outlier factor (LOF) algorithm. The rigorous empirical evaluation, conducted on pertinent real-world datasets, leverages classifiers based on long short-term memory (LSTM), convolutional neural networks (CNN), and transformer architectures to pinpoint adversarial incursions. The results underscore the superiority of our proposed technique in comparison to recent state-of-the-art methods, namely DISP and FGWS, achieving an impressive F1 detection accuracy rate of up to 94.8%.
Related Content
Sharon L. Burton.
© 2024.
25 pages.
|
Laura Ann Jones, Ian McAndrew.
© 2024.
24 pages.
|
Olayinka Creighton-Randall.
© 2024.
14 pages.
|
Stacey L. Morin.
© 2024.
11 pages.
|
N. Nagashri, L. Archana, Ramya Raghavan.
© 2024.
22 pages.
|
Esther Gani, Foluso Ayeni, Victor Mbarika, Abdullahi I. Musa, Oneurine Ngwa.
© 2024.
25 pages.
|
Sia Gholami, Marwan Omar.
© 2024.
18 pages.
|
|
|