Data Leverage References

← Back to browse

AI models collapse when trained on recursively generated data

2024 article shumailov2024modelcollapse Needs review - 1 field differ

Fields with differences: venue. Compare local vs external BibTeX below.

Authors
Ilia Shumailov, Zakhar Shumaylov, Yiren Zhao, Nicolas Papernot, Ross Anderson, Yarin Gal
Venue
Nature
Abstract
Landmark study showing that indiscriminate use of model-generated content in training causes irreversible defects in resulting models, where tails of original content distribution disappear. Model collapse is a degenerative learning process where models forget improbable events over time. Demonstrates this across LLMs, VAEs, and GMMs.

BibTeX

Local Entry
@article{shumailov2024modelcollapse,
  title = {AI models collapse when trained on recursively generated data},
  author = {Ilia Shumailov and Zakhar Shumaylov and Yiren Zhao and Nicolas Papernot and Ross Anderson and Yarin Gal},
  year = {2024},
  journal = {Nature},
  url = {https://www.nature.com/articles/s41586-024-07566-y},
  abstract = {Landmark study showing that indiscriminate use of model-generated content in training causes irreversible defects in resulting models, where tails of original content distribution disappear. Model collapse is a degenerative learning process where models forget improbable events over time. Demonstrates this across LLMs, VAEs, and GMMs.}
}
From AUTO:DBLP
@article{shumailov2024modelcollapse,
  title = {AI models collapse when trained on recursively generated data.},
  author = {Ilia Shumailov and Zakhar Shumaylov and Yiren Zhao and Nicolas Papernot and Ross J. Anderson and Yarin Gal},
  year = {2024},
  journal = {Nat.},
  doi = {10.1038/S41586-024-07566-Y}
}