Supervised Learning
Also known as: supervised training, labeled learning
What is Supervised Learning?
Supervised learning is a machine learning paradigm where a model is trained on labeled data, meaning each input comes paired with the correct output. The model learns to map inputs to outputs by minimizing the difference between its predictions and the true labels. It is the oldest and most intuitive form of machine learning: you show the system thousands of examples with answers, and it learns the pattern.
How It Works
In a supervised learning setup, a training dataset consists of input-output pairs. For image classification, the input might be a photograph and the label is the category (cat, dog, car). For text, the input could be a sentence and the label a sentiment score. The model processes each input, produces a prediction, compares it to the true label via a loss function, and then adjusts its internal parameters through backpropagation to reduce future errors. This cycle repeats across millions of examples until the model generalizes well to unseen data.
Supervised Learning in the LLM Era
While large language models are predominantly trained with self-supervised learning during pre-training, supervised learning remains critical in the fine-tuning phase. Instruction tuning, where models learn to follow human instructions, relies on curated datasets of prompt-response pairs created by human annotators. Supervised fine-tuning (SFT) is typically the first step after pre-training, before reinforcement learning from human feedback further refines the model’s behavior.
Strengths and Limitations
Supervised learning excels when high-quality labeled data is available and the task is well-defined. It powers production systems in medical imaging, fraud detection, speech recognition, and countless other domains. Its primary limitation is the dependency on labeled data, which is expensive and time-consuming to create at scale. This bottleneck is precisely what motivated the shift toward self-supervised approaches that can learn from unlabeled data.
Related Reading
- Deep Learning - The broader paradigm that powers modern supervised learning
- Reinforcement Learning - An alternative training paradigm using rewards instead of labels
- Backpropagation - The algorithm that makes supervised learning work