Ch.01
Supervised, Unsupervised, and Self-Supervised Learning
Machine learning is often divided into supervised, unsupervised, and self-supervised learning depending on how data is used. Supervised learning is like studying with an answer key; unsupervised learning is like finding patterns and grouping similar items without labels; self-supervised learning is like masking part of the data and learning by predicting the missing part. This chapter summarizes the core ideas, math, and real-world use of these three paradigms so you can build a solid base for the algorithms covered later.
ML diagram by chapter
Select a chapter to see its diagram below. View the machine learning flow at a glance.
Three learning paradigms: supervised (input–label pairs), unsupervised (no label), self-supervised (self-created target).
Supervised: input x and label y come in pairs
When (x, y) pairs are given in order, the model learns the rule
Unsupervised: only input x (no label y)
There is no y (label), only x. Some x blink on and off → the model still finds structure and clusters
Self-supervised: mask part of the data and predict the gap
e.g. fill in the blank → representation learning (BERT, etc.)
Three Ways of Learning: Supervised, Unsupervised, Self-Supervised
- Label
- SupervisedYes ()
- UnsupervisedNo
- Self-SupervisedSelf-created target
- Goal
- SupervisedPredict (classification/regression)
- UnsupervisedStructure, clusters, dimensionality reduction
- Self-SupervisedRepresentation learning
- Examples
- SupervisedKNN, linear/logistic regression
- UnsupervisedK-Means, dimension reduction
- Self-SupervisedBERT, contrastive learning
| Supervised | Unsupervised | Self-Supervised | |
|---|---|---|---|
| Label | Yes () | No | Self-created target |
| Goal | Predict (classification/regression) | Structure, clusters, dimensionality reduction | Representation learning |
| Examples | KNN, linear/logistic regression | K-Means, dimension reduction | BERT, contrastive learning |