Research

Diffusion Models Outperform Autoregressive AI in Data-Constrained Settings

A Carnegie Mellon University paper found diffusion models are more robust to data repetition with sufficient compute.

Olivia Sharp 1 min read 581 views
Free
New research from Carnegie Mellon University suggests diffusion models may offer a more efficient path for AI as training data becomes scarce.

A new paper from researchers at Carnegie Mellon University, published on September 22, 2025, addresses a fundamental question of AI model architecture and efficiency. The study found that diffusion models can outperform the more common autoregressive models, the basis for systems like GPT, in data-constrained settings, provided they are trained with sufficient compute.

## The Data Bottleneck

The research is motivated by the growing recognition that the AI field is approaching the limits of high-quality training data available on the internet. The paper sought to answer the question: "how can we trade off more compute for less data?"

Archive Access

This article is older than 24 hours. Create a free account to access our 7-day archive.

Share this article

Related Articles