More on Model Collapse

“architecture. In addition to being home to some of the world’s largest populations of black @-@ tailed jackrabbits, white @-@ tailed jackrabbits, blue @-@ tailed jackrabbits, red @-@ tailed jackrabbits, yellow @-.”

That’s a large language model trained on human input from Wikipedia then trained on its own output for nine iterations.

“For now, our store of human-generated data is large enough that current AI models won’t collapse overnight,” the researchers noted. “But to avoid a future where they do, AI developers will need to take more care about what they choose to feed into their systems. This doesn’t mean doing away with synthetic data entirely, but it does mean it will need to be better designed if models built on it are to work as intended.”