Skip to main content
Back to Glossary
Techniques

Data Augmentation

Techniques for artificially expanding training datasets by creating modified versions of existing data, like rotating images or paraphrasing text.


Why this matters

More training data usually means better AI models, but collecting data is expensive and time-consuming. Data augmentation is a clever shortcut. Take the examples you already have and create variations. Flip an image horizontally, add some noise, adjust the brightness. Now you have multiple training examples from one original. It's getting more value from data you've already paid for.

The technique has been standard practice in computer vision for years. Random crops, rotations, color shifts, these simple transformations can significantly improve model performance. The key insight is that a cat rotated 15 degrees is still a cat. By training on these variations, models learn to recognize objects regardless of minor differences in how they appear.

For text and language, augmentation is trickier but still useful. You can swap synonyms, shuffle sentence order, translate back and forth between languages, or use AI to generate paraphrases. The goal is the same: teach the model that minor variations don't change the underlying meaning. A customer complaint worded differently is still a complaint.

The technique does have limits. Augmentation can't add truly new information to your dataset. If all your cat photos are from the same angle, no amount of rotation will teach the model what a cat looks from below. And aggressive augmentation can introduce unrealistic examples that hurt performance. Like most things in AI, it requires judgment about what transformations actually make sense for your specific problem.

Related Terms

More in Techniques