AI

‘Model collapse’: Scientists warn against letting AI eat its own tail

Comment

Ouroboros
Image Credits: mariaflaya / Getty Images

When you see the mythical Ouroboros, it’s perfectly logical to think, “Well, that won’t last.” A potent symbol — swallowing your own tail — but difficult in practice. It may be the case for AI as well, which, according to a new study, may be at risk of “model collapse” after a few rounds of being trained on data it generated itself.

In a paper published in Nature, British and Canadian researchers led by Ilia Shumailov at Oxford show that today’s machine learning models are fundamentally vulnerable to a syndrome they call “model collapse.” As they write in the paper’s introduction:

We discover that indiscriminately learning from data produced by other models causes “model collapse” — a degenerative process whereby, over time, models forget the true underlying data distribution …

How does this happen, and why? The process is actually quite easy to understand.

AI models are pattern-matching systems at heart: They learn patterns in their training data, then match prompts to those patterns, filling in the most likely next dots on the line. Whether you ask, “What’s a good snickerdoodle recipe?” or “List the U.S. presidents in order of age at inauguration,” the model is basically just returning the most likely continuation of that series of words. (It’s different for image generators, but similar in many ways.)

But the thing is, models gravitate toward the most common output. It won’t give you a controversial snickerdoodle recipe but the most popular, ordinary one. And if you ask an image generator to make a picture of a dog, it won’t give you a rare breed it only saw two pictures of in its training data; you’ll probably get a golden retriever or a Lab.

Now, combine these two things with the fact that the web is being overrun by AI-generated content and that new AI models are likely to be ingesting and training on that content. That means they’re going to see a lot of goldens!

And once they’ve trained on this proliferation of goldens (or middle-of-the road blogspam, or fake faces, or generated songs), that is their new ground truth. They will think that 90% of dogs really are goldens, and therefore when asked to generate a dog, they will raise the proportion of goldens even higher — until they basically have lost track of what dogs are at all.

This wonderful illustration from Nature’s accompanying commentary article shows the process visually:

Image Credits: Nature

A similar thing happens with language models and others that, essentially, favor the most common data in their training set for answers — which, to be clear, is usually the right thing to do. It’s not really a problem until it meets up with the ocean of chum that is the public web right now.

Basically, if the models continue eating each other’s data, perhaps without even knowing it, they’ll progressively get weirder and dumber until they collapse. The researchers provide numerous examples and mitigation methods, but they go so far as to call model collapse “inevitable,” at least in theory.

Though it may not play out as the experiments they ran show it, the possibility should scare anyone in the AI space. Diversity and depth of training data is increasingly considered the single most important factor in the quality of a model. If you run out of data, but generating more risks model collapse, does that fundamentally limit today’s AI? If it does begin to happen, how will we know? And is there anything we can do to forestall or mitigate the problem?

The answer to the last question at least is probably yes, although that should not alleviate our concerns.

Qualitative and quantitative benchmarks of data sourcing and variety would help, but we’re far from standardizing those. Watermarks of AI-generated data would help other AIs avoid it, but so far no one has found a suitable way to mark imagery that way (well … I did).

In fact, companies may be disincentivized from sharing this kind of information, and instead hoard all the hyper-valuable original and human-generated data they can, retaining what Shumailov et al. call their “first mover advantage.”

[Model collapse] must be taken seriously if we are to sustain the benefits of training from large-scale data scraped from the web. Indeed, the value of data collected about genuine human interactions with systems will be increasingly valuable in the presence of LLM-generated content in data crawled from the Internet.

… [I]t may become increasingly difficult to train newer versions of LLMs without access to data that were crawled from the Internet before the mass adoption of the technology or direct access to data generated by humans at scale.

Add it to the pile of potentially catastrophic challenges for AI models — and arguments against today’s methods producing tomorrow’s superintelligence.

More TechCrunch

Supermaven, an AI coding assistant, has raised $12 million in a funding round that had participation from OpenAI and Perplexity co-founders.

AI coding assistant Supermaven raises cash from OpenAI and Perplexity co-founders

Arjun Vora and Tito Goldstein were working on the corporate side of Uber when they realized that HR software largely wasn’t built to manage hourly staff. Many hourly workers lacked…

TeamBridge, founded by former Uber execs, raises $28M to build HR software for hourly workers

The US Food and Drug Administration Monday published approval for sleep apnea detection on the Apple Watch Series 9, Series 10, and Watch Ultra 2. The green light comes four…

Apple Watch sleep apnea detection gets FDA approval

Featured Article

Apple AirPods 4 with Active Noise Cancellation review

I can’t recall another consumer electronics product category becoming a commodity as quickly as Bluetooth earbuds. Apple’s AirPods played a key role in that growth, of course, recapturing a kind of excitement not seen in consumer music tech since the original iPod. AirPods’ fundamentals haven’t changed much in the eight…

Apple AirPods 4 with Active Noise Cancellation review

Myntra, India’s largest fashion e-commerce platform, is trialling a four-hour delivery service in four Indian cities, two sources familiar with the matter told TechCrunch, a dramatic acceleration from its standard…

Myntra bets on 4-hour delivery amid India’s quick commerce boom

AWS today announced that it is transitioning OpenSearch, its open source fork of the popular Elasticsearch search and analytics engine, to the Linux Foundation with the launch of the very…

AWS brings OpenSearch under the Linux Foundation umbrella

Insight Partners is reportedly on the cusp of on more than $10 billion in capital commitments for its 13th fund, per the FT.  The FT report notes that two of…

Insight Partners is closing in on a whopping $10B+ new fund

The Port of Seattle released a statement Friday confirming that it was targeted by a ransomware attack. The attack occurred on August 24, with the Port (which also operates the…

Port of Seattle shares ransomware attack details

A decade after the wildly popular game Flappy Bird disappeared, an organization calling itself The Flappy Bird Foundation announced plans to “re-hatch the official Flappy Bird® game.” But this morning,…

Flappy Bird’s creator disavows ‘official’ new version of the game

Platforms to connect apps that wouldn’t normally talk to each other have been around for a minute (see: Zapier). But they have not gotten dramatically simpler to use if you’re…

DryMerge promises to connect apps that normally don’t talk to each other — and when it works, it’s great

Featured Article

Cohere co-founder Nick Frosst’s indie band, Good Kid, is almost as successful as his AI company

Nick Frosst, the co-founder of $5.5 billion Canadian AI startup Cohere, has been a musician his whole life. He told TechCrunch that once he started singing, he never shut up. That’s still true today. In addition to his full-time job at Cohere, Frosst is also the front man of Good…

Cohere co-founder Nick Frosst’s indie band, Good Kid, is almost as successful as his AI company

Blockchain technology is all about decentralization and virtualization. So it’s a little ironic that humans love to come together in person at big blockchain events. Such was the case last…

A walk through the crypto jungle at Korea Blockchain Week

I have a guilty pleasure, and it’s not that I just rewatched “Glee” in its entirety (yes, even the awful later seasons), or that I have read an ungodly amount…

The LinkedIn games are fun, actually

It’s looking increasingly likely that OpenAI will soon alter its complex corporate structure. Reports earlier this week suggested that the AI company was in talks to raise $6.5 billion at…

OpenAI could shake up its nonprofit structure next year

Fusion startups have raised $7.1 billion to date, with the majority of it going to a handful of companies. 

Every fusion startup that has raised over $300M

Netflix has never quite cracked the talk show formula, but maybe it can borrow an existing hit from YouTube. According to Bloomberg, the streamer is in talks with BuzzFeed to…

‘Hot Ones’ could add some heat to Netflix’s live lineup

Alex Parmley has been thinking about building his latest company, ORNG, since he was working on his last company, Phood.  Launched in 2018, Phood was a payments app that let…

Why ORNG’s founder pivoted from college food ordering to real-time money transfer

Lawyers representing Sam Bankman-Fried, the FTX CEO and co-founder who was convicted of fraud and money laundering late last year, are seeking a new trial. Following crypto exchange FTX’s collapse,…

Sam Bankman-Fried appeals conviction, criticizes judge’s ‘unbalanced’ decisions

OpenAI this week unveiled a preview of OpenAI o1, also known as Strawberry. The company claims that o1 can more effectively reason through math and science, as well as fact-check…

OpenAI previews its new Strawberry model

There’s something oddly refreshing about starting the day by solving the Wordle. According to DeepWell DTx, there’s a scientific explanation for why our brains might feel just a bit better…

DeepWell DTx receives FDA clearance for its therapeutic video game developer tools

Soundiiz is a free third-party tool that builds portability tools through existing APIs and acts as a translator between the services.

These two friends built a simple tool to transfer playlists between Apple Music and Spotify, and it works great

In early 2018, VC Mike Moritz wrote in the FT that “Silicon Valley would be wise to follow China’s lead,” noting the pace of work at tech companies was “furious”…

This is how bad China’s startup scene looks now

Fei-Fei Li, the Stanford professor many deem the “Godmother of AI,” has raised $230 million for her new startup, World Labs, from backers including Andreessen Horowitz, NEA, and Radical Ventures.…

Fei-Fei Li’s World Labs comes out of stealth with $230M in funding

Bolt says it has settled its long-standing lawsuit with its investor Activant Capital. One-click payments startup Bolt is settling the suit by buying out the investor’s stake “after which Activant…

Fintech Bolt is buying out the investor suing over Ryan Breslow’s $30M loan

The rise of neobanks has been fascinating to witness, as a number of companies in recent years have grown from merely challenging traditional banks to being massive players in and…

Dave and Varo Bank execs are coming to TechCrunch Disrupt 2024

OpenAI released its new o1 models on Thursday, giving ChatGPT users their first chance to try AI models that pause to “think” before they answer. There’s been a lot of…

First impressions of OpenAI o1: An AI designed to overthink it

Featured Article

Investors rebel as TuSimple pivots from self-driving trucks to AI gaming

TuSimple, once a buzzy startup considered a leader in self-driving trucks, is trying to move its assets to China to fund a new AI-generated animation and video game business. The pivot has not only puzzled and enraged several shareholders, but also threatens to pull the company back into a legal…

Investors rebel as TuSimple pivots from self-driving trucks to AI gaming

Welcome to Startups Weekly — your weekly recap of everything you can’t miss from the world of startups. Want it in your inbox every Friday? Sign up here. This week…

Shrinking teams, warped views, and risk aversion in this week’s startup news

Silicon Valley startup accelerator Y Combinator will expand the number of cohorts it runs each year from two to four starting in 2025, Bloomberg reported Thursday, and TechCrunch confirmed today.…

Y Combinator expanding to four cohorts a year in 2025

Telegram has had a tough few weeks. The messaging app’s founder, Pavel Durov, was arrested in late August and later released on a €5 million bail in France, charged with…

Telegram CEO Durov’s arrest hasn’t dampened enthusiasm for its TON blockchain