Data Leverage References

← Back to browse

Deep Double Descent: Where Bigger Models and More Data Hurt

2020 inproceedings nakkiran2020doubledescent Not yet verified
Authors
Preetum Nakkiran, Gal Kaplun, Yamini Bansal, Tristan Yang, Boaz Barak, Ilya Sutskever
Venue
ICLR 2020
Abstract
Demonstrates that double descent occurs across model size, training epochs, and dataset size in modern deep networks. Introduces effective model complexity to unify these phenomena and shows regimes where more data hurts.