The IRMA Community
Newsletters
Research IRM
Click a keyword to search titles using our InfoSci-OnDemand powered search:
|
Adversarial Attacks on Graph Neural Network: Techniques and Countermeasures
Abstract
Graph neural networks (GNNs) are a useful tool for analyzing graph-based data in areas like social networks, molecular chemistry, and recommendation systems. Adversarial attacks on GNNs include introducing malicious perturbations that manipulate the model's predictions without being detected. These attacks can be structural or feature-based depending on whether the attacker modifies the graph's topology or node/edge features. To defend against adversarial attacks, researchers have proposed countermeasures like robust training, adversarial training, and defense mechanisms that identify and correct adversarial examples. These methods aim to improve the model's generalization capabilities, enforce regularization, and incorporate defense mechanisms into the model architecture to improve its robustness against attacks. This chapter offers an overview of recent advances in adversarial attacks on GNNs, including attack methods, evaluation metrics, and their impact on model performance.
Related Content
Vinod Kumar, Himanshu Prajapati, Sasikala Ponnusamy.
© 2023.
18 pages.
|
Sougatamoy Biswas.
© 2023.
14 pages.
|
Ganga Devi S. V. S..
© 2023.
10 pages.
|
Gotam Singh Lalotra, Ashok Sharma, Barun Kumar Bhatti, Suresh Singh.
© 2023.
15 pages.
|
Nimish Kumar, Himanshu Verma, Yogesh Kumar Sharma.
© 2023.
16 pages.
|
R. Soujanya, Ravi Mohan Sharma, Manish Manish Maheshwari, Divya Prakash Shrivastava.
© 2023.
12 pages.
|
Nimish Kumar, Himanshu Verma, Yogesh Kumar Sharma.
© 2023.
22 pages.
|
|
|