March 19, 2024, 4:11 a.m. | Tingting Tang, Yue Niu, Salman Avestimehr, Murali Annavaram

cs.CR updates on arXiv.org arxiv.org

arXiv:2403.10995v1 Announce Type: cross
Abstract: Graph neural networks (GNNs) play a key role in learning representations from graph-structured data and are demonstrated to be useful in many applications. However, the GNN training pipeline has been shown to be vulnerable to node feature leakage and edge extraction attacks. This paper investigates a scenario where an attacker aims to recover private edge information from a trained GNN model. Previous studies have employed differential privacy (DP) to add noise directly to the adjacency …

applications arxiv attacks cs.ai cs.cr cs.lg cs.si data edge extraction feature graph key networks neural networks node pipeline play private role structured data training value vulnerable

CyberSOC Technical Lead

@ Integrity360 | Sandyford, Dublin, Ireland

Cyber Security Strategy Consultant

@ Capco | New York City

Cyber Security Senior Consultant

@ Capco | Chicago, IL

Sr. Product Manager

@ MixMode | Remote, US

Corporate Intern - Information Security (Year Round)

@ Associated Bank | US WI Remote

Senior Offensive Security Engineer

@ CoStar Group | US-DC Washington, DC