Author or contributor
Resource type

Understanding deep learning requires rethinking generalization

Resource type
Authors/contributors
Title
Understanding deep learning requires rethinking generalization
Abstract
Despite their massive size, successful deep artificial neural networks can exhibit a remarkably small difference between training and test performance. Conventional wisdom attributes small generalization error either to properties of the model family, or to the regularization techniques used during training. Through extensive systematic experiments, we show how these traditional approaches fail to explain why large neural networks generalize well in practice. Specifically, our experiments establish that state-of-the-art convolutional networks for image classification trained with stochastic gradient methods easily fit a random labeling of the training data. This phenomenon is qualitatively unaffected by explicit regularization, and occurs even if we replace the true images by completely unstructured random noise. We corroborate these experimental findings with a theoretical construction showing that simple depth two neural networks already have perfect finite sample expressivity as soon as the number of parameters exceeds the number of data points as it usually does in practice. We interpret our experimental findings by comparison with traditional models.
Publication
arXiv:1611.03530 [cs]
Date
2017-02-26
Accessed
2019-11-22T20:11:42Z
Library Catalog
Extra
ZSCC: NoCitationData[s0] arXiv: 1611.03530
Notes
Comment: Published in ICLR 2017
Citation
Zhang, C., Bengio, S., Hardt, M., Recht, B., & Vinyals, O. (2017). Understanding deep learning requires rethinking generalization. ArXiv:1611.03530 [Cs]. Retrieved from http://arxiv.org/abs/1611.03530
Attachment
Processing time: 0.03 seconds

Graph of references

(from Zotero to Gephi via Zotnet with this script)
Graph of references