June 6, 2022, 1:20 a.m. | Héber H. Arcolezi, Jean-François Couchot, Denis Renaud, Bechara Al Bouna, Xiaokui Xiao

cs.CR updates on arXiv.org arxiv.org

This paper investigates the problem of forecasting multivariate aggregated
human mobility while preserving the privacy of the individuals concerned.
Differential privacy, a state-of-the-art formal notion, has been used as the
privacy guarantee in two different and independent steps when training deep
learning models. On one hand, we considered \textit{gradient perturbation},
which uses the differentially private stochastic gradient descent algorithm to
guarantee the privacy of each time series sample in the learning stage. On the
other hand, we considered \textit{input perturbation}, …

deep learning human input lg mobility

Sr. Cloud Security Engineer

@ BLOCKCHAINS | USA - Remote

Network Security (SDWAN: Velocloud) Infrastructure Lead

@ Sopra Steria | Noida, Uttar Pradesh, India

Senior Python Engineer, Cloud Security

@ Darktrace | Cambridge

Senior Security Consultant

@ Nokia | United States

Manager, Threat Operations

@ Ivanti | United States, Remote

Lead Cybersecurity Architect - Threat Modeling | AWS Cloud Security

@ JPMorgan Chase & Co. | Columbus, OH, United States