Using Trusted Data to Train Deep Networks on Labels Corrupted by Severe Noise

February 14, 2018 ยท Entered Twilight ยท ๐Ÿ› Neural Information Processing Systems

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 7.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: CIFAR, LICENSE, MNIST, README.md, SST, Twitter, glc_plots_figure.png, glc_vision_results.png

Authors Dan Hendrycks, Mantas Mazeika, Duncan Wilson, Kevin Gimpel arXiv ID 1802.05300 Category cs.LG: Machine Learning Cross-listed cs.CL, cs.CV, cs.NE Citations 597 Venue Neural Information Processing Systems Repository https://github.com/mmazeika/glc โญ 88 Last Checked 1 month ago
Abstract
The growing importance of massive datasets used for deep learning makes robustness to label noise a critical property for classifiers to have. Sources of label noise include automatic labeling, non-expert labeling, and label corruption by data poisoning adversaries. Numerous previous works assume that no source of labels can be trusted. We relax this assumption and assume that a small subset of the training data is trusted. This enables substantial label corruption robustness performance gains. In addition, particularly severe label noise can be combated by using a set of trusted data with clean labels. We utilize trusted data by proposing a loss correction technique that utilizes trusted examples in a data-efficient manner to mitigate the effects of label noise on deep neural network classifiers. Across vision and natural language processing tasks, we experiment with various label noises at several strengths, and show that our method significantly outperforms existing methods.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning