Data Leverage References

← Back to browse

Tag: clipped (3 references)

A Neural Scaling Law from the Dimension of the Data Manifold 2020 misc

Utkarsh Sharma, J. Kaplan

When data is plentiful, the loss achieved by well-trained neural networks scales as a power-law $L \propto N^{-\alpha}$ in the number of network parameters $N$. This empirical scaling law holds for a wide variety of data modalities, and may persist over many orders of magnitude. The scaling law can be explained if neural models are effectively just performing regression on a data manifold of intrinsic dimension $d$. This simple theory predicts that the scaling exponents $\alpha \approx 4/d$ for cross-entropy and mean-squared error losses. We confirm the theory by independently measuring the intrinsic dimension and the scaling exponents in a teacher/student framework, where we can study a variety of $d$ and $\alpha$ by dialing the properties of random teacher networks. We also test the theory with CNN image classifiers on several datasets and with GPT-type language models.

Turkers, Scholars, "Arafat" and "Peace": Cultural Communities and Algorithmic Gold Standards 2015 misc

Shilad Sen, Margaret E. Giesel, Rebecca Gold, Benjamin Hillmann, Matt Lesicko, Samuel Naden, Jesse Russell, Zixiao Wang, Brent J. Hecht

Recommender systems 1997 misc

Paul Resnick, H. Varian