Will We Run Out of Data?

Navigating the Impending AI Data Crisis
Publish Date
June 11, 2024
AI Expertise and Resources
Ani Bisaria

Artificial intelligence (AI) systems, such as ChatGPT, are on the cusp of encountering a critical challenge: the depletion of high-quality human-generated text data. A recent study by Epoch AI forecasts that the supply of publicly available training data for AI language models could be exhausted between 2026 and 2032. This impending data scarcity threatens the scaling and performance enhancement of AI models, highlighting the urgent need for innovative solutions (Masanet et al., 2020; Jones, 2018).

Addressing Immediate Data Limitations

Currently, tech giants like OpenAI and Google are tackling these data constraints by sourcing high-quality data. This often involves purchasing data access from platforms like Reddit and various news media outlets. In the short term, these companies are also optimizing the use of available datasets through advanced techniques such as data augmentation, which transforms existing data to create new training examples (Qiu, 2020).

For instance, OpenAI's dataset has been growing by 2.5 times annually, while the compute requirements for model training have increased fourfold each year (Van Heddeghem et al., 2014). However, repeatedly training models on the same data to extract maximum value poses the risk of overfitting and reducing generalizability, as evidenced by Meta's Llama 3 model, which was trained on 15 trillion tokens (Koomey et al., 2011).

Long-Term Challenges and Strategies

As the supply of fresh, high-quality human-generated text dwindles, AI developers must explore alternative data sources and methods. One controversial approach involves leveraging sensitive private data from emails, text messages, and other private communications. This raises significant privacy and security concerns, making it neither a sustainable nor ethical long-term solution (Shehabi et al., 2016).

Another potential strategy is the generation of synthetic data using AI technologies. However, this approach risks 'model collapse,' where the model's performance degrades over time due to the repetitive and potentially inaccurate nature of synthetic data (Patterson & Rumsey, 2003).

Insights and Projections from Epoch AI Study

The Epoch AI study provides critical insights and projections regarding the future of AI training data. It projects that high-quality text data will be exhausted between 2026 and 2032, depending on factors such as data overtraining rates and advancements in data efficiency (Masanet et al., 2013). For example, the quality-adjusted data stock is estimated at 320 trillion tokens. Current AI models, like those developed by OpenAI, are trained on datasets that grow approximately 2.5 times per year, whereas the computing power required for training is increasing at a rate of four times per year (Yahoo Finance, 2023).

Historically, the size of training datasets has been increasing by about 0.38 orders of magnitude annually, translating to roughly 2.4 times per year. This rapid growth may not be sustainable long-term, as only up to 40% of web data can be used as training data after deduplication without significantly compromising model performance (Masanet et al., 2020).

Exploring Alternative Data Sources

As human-generated text data becomes scarcer, AI developers may explore several alternative sources and methods to sustain performance. One approach involves using AI-generated data. While synthetic data can be useful in specific domains like mathematics and programming, its effectiveness for general-purpose natural language processing models is limited due to issues like information loss and lack of diversity (Qiu, 2020).

Another approach is incorporating data from other modalities, through AI technologies such as image recognition capabilities and video analysis. While this can temporarily alleviate the text data bottleneck, it may not fully compensate for the lack of text data (Jones, 2018). Additionally, tapping into non-indexed data from social media, private messaging apps, and other sources could provide additional machine learning training data. However, this raises significant ethical and privacy concerns (Van Heddeghem et al., 2014).

Future Outlook

The AI community faces a critical juncture as it grapples with impending data scarcity. Developing robust methods for generating and utilizing synthetic AI data, as well as transferring knowledge from other domains, could help address data shortages. However, these methods require further research and refinement to ensure their effectiveness (Koomey et al., 2011).

Advancements in data efficiency techniques, such as better data filtering and augmentation methods, can help maximize the utility of existing data stocks. Moreover, addressing the ethical implications of using private and sensitive data for AI training is crucial. Developing policies and frameworks that prioritize user privacy and data security will be essential for sustainable AI development (Shehabi et al., 2016).

In conclusion, the potential exhaustion of high-quality human-generated text data presents a significant challenge for the future of AI development. While innovative approaches such as synthetic data generation, multimodal learning, and improved data efficiency offer potential solutions, the urgency of developing these strategies is paramount. By understanding and addressing these challenges, the AI community can work towards creating more resilient and adaptable models that continue to push the boundaries of what artificial intelligence can achieve.


  • Jones, N. (2018). How to stop data centres from gobbling up the world’s electricity. Nature, 561(7722), 163-166. https://doi.org/10.1038/d41586-018-06610-y
  • Koomey, J. G., et al. (2011). Implications of historical trends in the electrical efficiency of computing. IEEE Annals of the History of Computing, 33(3), 46-54. https://doi.org/10.1109/MAHC.2010.28
  • Masanet, E., et al. (2020). Recalibrating global data center energy-use estimates. Science, 367(6481), 984-986. https://doi.org/10.1126/science.aba3758
  • Masanet, E., et al. (2013). The energy efficiency potential of cloud-based software: A U.S. case study. Environmental Research Letters, 8(3), 035018. https://doi.org/10.1088/1748-9326/8/3/035018
  • Patterson, M. K., & Rumsey, A. W. (2003). Effective thermal management in data centers. Intel Technology Journal, 7(1), 17-26.
  • Qiu, J. (2020). Big data’s big potential for HR. Harvard Business Review. https://hbr.org/2020/05/big-datas-big-potential-for-hr

Join the 40 other companies that trust The Matchbox to distinguish their brand in a crowded market.