张浩彬 (2024-07-29 13:18):
#paper DOI: https://doi.org/10.1038/s41586-024-07566-y ,AI models collapse when trained on recursively generated data。Nature关于大模型合成语料的探讨文章,讨论了在训练数据中,合成语料的加入(可能是被动,由于现有网络资料已经大量的大模型合成语料),导致模型崩溃的问题。当然,合成语料的使用易燃是大模型的训练的有效方式,但是要做好对合成语料的筛选工作
IF:50.500Q1 Nature, 2024-07-24T15:01:51. DOI: 10.1038/s41586-024-07566-y
AI models collapse when trained on recursively generated data
翻译
Abstract:
AbstractStable diffusion revolutionized image creation from descriptive text. GPT-2 (ref. 1), GPT-3(.5) (ref. 2) and GPT-4 (ref. 3) demonstrated high performance across a variety of language tasks. ChatGPT introduced such language models to the public. It is now clear that generative artificial intelligence (AI) such as large language models (LLMs) is here to stay and will substantially change the ecosystem of online text and images. Here we consider what may happen to GPT-{n} once LLMs contribute much of the text found online. We find that indiscriminate use of model-generated content in training causes irreversible defects in the resulting models, in which tails of the original content distribution disappear. We refer to this effect as ‘model collapse’ and show that it can occur in LLMs as well as in variational autoencoders (VAEs) and Gaussian mixture models (GMMs). We build theoretical intuition behind the phenomenon and portray its ubiquity among all learned generative models. We demonstrate that it must be taken seriously if we are to sustain the benefits of training from large-scale data scraped from the web. Indeed, the value of data collected about genuine human interactions with systems will be increasingly valuable in the presence of LLM-generated content in data crawled from the Internet.
翻译
回到顶部