Skip to main content
Back to Glossary
Models

Chinchilla

DeepMind's language model that proved smaller, better-trained models outperform larger undertrained ones, reshaping how the field thinks about scaling.


What is Chinchilla?

Chinchilla is a 70 billion parameter language model from DeepMind, published in 2022. But its importance isn't the model itself. It's what Chinchilla proved about training efficiency. The Chinchilla paper showed that most large language models were undertrained. You could get better results with a smaller model trained on more data than a larger model trained on less.

The Chinchilla Scaling Laws

Before Chinchilla, the assumption was bigger is better. Build the largest model you can afford, train it until your budget runs out. Chinchilla upended this. The research showed that model size and training data should scale together. For a given compute budget, there's an optimal balance. Most models at the time were oversized and undertrained. GPT-3 had 175B parameters but was trained on only 300B tokens. Chinchilla suggested it should have seen far more data.

How Chinchilla Changed AI Development

The Chinchilla findings influenced every major model that came after. LLaMA explicitly followed Chinchilla scaling laws, training smaller models on much more data to great effect. The shift toward training efficiency over raw size traces directly back to this paper. It's why we now see capable 7B and 13B models that punch far above their weight.

Strengths and Legacy

Chinchilla's strength was scientific rigor. It gave the field clear guidelines for efficient training. The model itself performed comparably to much larger models at lower cost. The main limitation is that these scaling laws aren't universal. They apply to the pretraining phase but don't account for fine-tuning, where other dynamics apply. Still, Chinchilla remains one of the most influential papers in modern AI research.

Related Terms

More in Models