all InfoSec news
Releasing Graph Neural Networks with Differential Privacy Guarantees. (arXiv:2109.08907v2 [cs.LG] UPDATED)
cs.CR updates on arXiv.org arxiv.org
With the increasing popularity of graph neural networks (GNNs) in several
sensitive applications like healthcare and medicine, concerns have been raised
over the privacy aspects of trained GNNs. More notably, GNNs are vulnerable to
privacy attacks, such as membership inference attacks, even if only black-box
access to the trained model is granted. We propose PrivGNN, a
privacy-preserving framework for releasing GNN models in a centralized setting.
Assuming an access to a public unlabeled graph, PrivGNN provides a framework to
release …
access applications attacks box differential privacy graph healthcare medicine networks neural networks privacy sensitive vulnerable