Zubnet AILearnWiki › Data Augmentation
Training

Data Augmentation

Techniques that artificially expand a training dataset by creating modified versions of existing examples. For images: flipping, rotating, cropping, color shifting. For text: paraphrasing, back-translation, synonym substitution. For audio: speed changes, noise injection. The goal is to teach the model invariances — a cat is a cat whether the image is flipped, darkened, or cropped.

Why it matters

Data augmentation is the cheapest way to improve model performance when you have limited data. It reduces overfitting by showing the model many variations of each example, teaching it to focus on essential features rather than superficial details. In computer vision, augmentation routinely provides 2–5% accuracy improvements for free.

Deep Dive

The key principle: augmentations must preserve the label. Flipping a cat image horizontally still shows a cat (valid augmentation). Flipping a "turn left" sign makes it a "turn right" sign (invalid augmentation). Choosing appropriate augmentations requires understanding what invariances matter for your task.

Modern Augmentation

AutoAugment and its successors (RandAugment, TrivialAugment) learn or randomize augmentation policies instead of hand-designing them. Cutout/CutMix randomly masks or mixes patches from different images. MixUp interpolates between pairs of examples, creating synthetic training points that smooth decision boundaries. These techniques are now standard in vision training pipelines.

AI-Powered Augmentation

With generative models, augmentation goes beyond geometric transforms. You can use LLMs to paraphrase text training data, use diffusion models to generate variant images, or use models to create entirely new training examples (synthetic data). The line between "augmentation" (modifying existing examples) and "synthetic data" (generating new examples) is blurring, and both are becoming essential parts of modern training pipelines.

Related Concepts

← All Terms
← Cursor Data Centers →