Researchers Warn We Could Run Out of Data to Train AI by 2026. What Then?

As artificial intelligence reaches the peak of its popularity, researchers have warned the industry might be running out of training data—the fuel that runs powerful AI systems. This could slow down the growth of AI models, especially large language models, and may even alter the trajectory of the AI revolution.

But why is a potential lack of data an issue, considering how much there is on the web? And is there a way to address the risk?

Why High-Quality Data Is Important for AI

We need a lot of data to train powerful, accurate, and high-quality AI algorithms. For instance, the algorithm powering ChatGPT was originally trained on 570 gigabytes of text data, or about 300 billion words.

Similarly, the Stable Diffusion algorithm (which is behind many AI image-generating apps) was trained on the LAION-5B dataset comprised of 5.8 billion image-text pairs. If an algorithm is trained on an insufficient amount of data, it will produce inaccurate or low-quality outputs.

The quality of the training data is also important. Low-quality data such as social media posts or blurry photographs are easy to source but aren’t sufficient to train high-performing AI models.

Text taken from social media platforms might be biased or prejudiced, or may include disinformation or illegal content which could be replicated by the model. For example, when Microsoft tried to train its AI bot using Twitter content, it learned to produce racist and misogynistic outputs.

This is why AI developers seek out high-quality content such as text from books, online articles, scientific papers, Wikipedia, and certain filtered web content. The Google Assistant was trained on 11,000 romance novels taken from self-publishing site Smashwords to make it more conversational.

Do We Have Enough Data?

The AI industry has been training AI systems on ever-larger datasets, which is why we now have high-performing models such as ChatGPT or DALL-E 3. At the same time, research shows online data stocks are growing much more slowly than datasets used to train AI.

In a paper published last year, a group of researchers predicted we will run out of high-quality text data before 2026 if current AI training trends continue. They also estimated low-quality language data will be exhausted sometime between 2030 and 2050, and low-quality image data between 2030 and 2060.

AI could contribute up to $15.7 trillion to the world economy by 2030, according to accounting and consulting group PwC. But running out of usable data could slow down its development.

Should We Be Worried?

While the above points might alarm some AI fans, the situation may not be as bad as it seems. There are many unknowns about how AI models will develop in the future, as well as a few ways to address the risk of data shortages.

One opportunity is for AI developers to improve algorithms so they use the data they already have more efficiently.

It’s likely in the coming years they will be able to train high-performing AI systems using less data, and possibly less computational power. This would also help reduce AI’s carbon footprint.

Another option is to use AI to create synthetic data to train systems. In other words, developers can simply generate the data they need, curated to suit their particular AI model.

Several projects are already using synthetic content, often sourced from data-generating services such as Mostly AI. This will become more common in the future.

Developers are also searching for content outside the free online space, such as that held by large publishers and offline repositories. Think about the millions of texts published before the internet. Made available digitally, they could provide a new source of data for AI projects.

News Corp, one of the world’s largest news content owners (which has much of its content behind a paywall) recently said it was negotiating content deals with AI developers. Such deals would force AI companies to pay for training data—whereas they have mostly scraped it off the internet for free so far.

Content creators have protested against the unauthorized use of their content to train AI models, with some suing companies such as Microsoft, OpenAI, and Stability AI. Being remunerated for their work may help restore some of the power imbalance that exists between creatives and AI companies.

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Image Credit: Emil Widlund / Unsplash

Rita Matulionyte
Rita Matulionytehttps://researchers.mq.edu.au/en/persons/rita-matulionyte
Rita Matulionyte is an international expert in intellectual property and information technology law, with a recent focus on legal regulation and governance of artificial intelligence technologies. She acquired her PhD from Albert Ludwig University of Freiburg/Max Planck Institute for intellectual property and competition law in 2010. Since then she has researched and lectured in universities in Japan, Germany, Lithuania, and Australia. To date, she has published around 40 peer-reviewed articles and book chapters, as well as a monograph (Applicable Law to Copyright Infringement: A Comparison of ALI and CLIP Principles). Rita has been invited to present in conferences in Europe, Japan, Hong Kong, and the United States, and has prepared reports for the European Commission, European Patent Office, and governments of South Korea and Lithuania.
RELATED
latest
Don't miss a trend
Get Hub delivered to your inbox

featured