Double descent
In statistics and machine learning, double descent is the phenomenon where a statistical model with a small number of parameters and a model with an extremely large number of parameters have a small error, but a model whose number of parameters is about the same as the number of data points used to train the model will have a large error.[1] It was discovered in 2019 when researchers were trying to reconcile the bias-variance tradeoff in classical statistics, which states that having too many parameters will yield an extremely large error, with the 2010s empirical observation of machine learning practitioners that the larger models are, the better they work.
Part of a series on |
Machine learning and data mining |
---|
![]() |
See also
References
- "Deep Double Descent". OpenAI. 2019-12-05. Retrieved 2022-08-12.
- Mikhail Belkin; Daniel Hsu; Ji Xu (2020). "Two Models of Double Descent for Weak Features". SIAM Journal on Mathematics of Data Science. 2 (4). doi:10.1137/20M1336072.
- Preetum Nakkiran; Gal Kaplun; Yamini Bansal; Tristan Yang; Boaz Barak; Ilya Sutskever (29 December 2021). "Deep double descent: where bigger models and more data hurt". Journal of Statistical Mechanics: Theory and Experiment. IOP Publishing Ltd and SISSA Medialab srl. 2021. arXiv:1912.02292. doi:10.1088/1742-5468/ac3a74.
- Song Mei; Andrea Montanari (April 2022). "The Generalization Error of Random Features Regression: Precise Asymptotics and the Double Descent Curve". Communications on Pure and Applied Mathematics. 75 (4). arXiv:1908.05355. doi:10.1002/cpa.22008.
- Xiangyu Chang; Yingcong Li; Samet Oymak; Christos Thrampoulidis (2021). "Provable Benefits of Overparameterization in Model Compression: From Double Descent to Pruning Neural Networks". Proceedings of the AAAI Conference on Artificial Intelligence. 35 (8). arXiv:2012.08749.
- Mikhail Belkin; Daniel Hsu; Siyuan Ma; Soumik Mandal (2019). "Reconciling modern machine-learning practice and the classical bias–variance trade-off". Proceedings of the National Academy of Sciences of the United States of America. 116 (32). doi:10.1073/pnas.1903070116.
- Marco Loog; Tom Viering; Alexander Mey; Jesse H. Krijthe; David M. J. Tax (2020). "A brief prehistory of double descent". Proceedings of the National Academy of Sciences of the United States of America. 117 (16). doi:10.1073/pnas.2001875117.
External links
- Brent Werness; Jared Wilber. "Double Descent: Part 1: A Visual Introduction".
- Brent Werness; Jared Wilber. "Double Descent: Part 2: A Mathematical Explanation".
This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.