May 28, 2024, 4:12 a.m. | Noga Bar, Tomer Koren, Raja Giryes

cs.CR updates on arXiv.org arxiv.org

arXiv:2102.12192v4 Announce Type: replace-cross
Abstract: Neural networks are widespread due to their powerful performance. However, they degrade in the presence of noisy labels at training time. Inspired by the setting of learning with expert advice, where multiplicative weight (MW) updates were recently shown to be robust to moderate data corruptions in expert advice, we propose to use MW for reweighting examples during neural networks optimization. We theoretically establish the convergence of our method when used with gradient descent and prove …

advice arxiv cs.cr cs.lg data expert network networks neural network neural networks noisy optimization performance presence training updates

Ingénieur Développement Logiciel IoT H/F

@ Socomec Group | Benfeld, Grand Est, France

Architecte Cloud – Lyon

@ Sopra Steria | Limonest, France

Senior Risk Operations Analyst

@ Visa | Austin, TX, United States

Military Orders Writer

@ Advanced Technology Leaders, Inc. | Ft Eisenhower, GA, US

Senior Golang Software Developer (f/m/d)

@ E.ON | Essen, DE

Senior Revenue Operations Analyst (Redwood City)

@ Anomali | Redwood City, CA