WTF Learning
Machine learning has so many words, so many ideas, it’s becoming increasing diffiucult to know what is going on. I have tried to capture some coherence here to aid discussions.
Some background terms
- Model: A machine learning model
- Task: the problem or objective a model is required to solve
- Skill: the ability of a model to perform a specific task
- Learning: the process by which a model improves its skill after gaining experience
- Pretext Task: a contrived task that may lead to learning on the actual task
- Target: the outcome expected for a given task or pretext task
- Sample/Example: a single unit of data observed in the learning process
- Dataset: collection of samples used to train or test a model
- Training: the process of increasing a models skill during learning
- Testing: the process of testing a models skill during learning
- Domain: specific area or field of knowledge from which the dataset is comes
- Label/Class: a categorical target associated with a sample in a classification task
- Classes: the set of labels in a classification task
- Knowledge: factual information, patterns, and relationships a model has learned
- Embedding Space:
- Attributes:
Data Distribution Contraints
These are constraints placed on the learning due to the dataset that is available.
- Supervised learning
dataset includes the task's target
- Semi-supervised learning
only a subset of the dataset includes the task's target
- Unsupervised learning
dataset does not include the task's target
- Self-supervised learning
target created from a pretext task
- Reinforcement learning
dataset only include task's target in last sample
Data Sampling Constraints
The contraints arise as a result of limitations on how we can sample the dataset.
- (?) Zero-Shot learning
use class attributes rather than label as target
- One-Shot learning
each class has one training sample
- Few-Shot learning
each class has fewer than six training samples
- Continual/Lifelong/Incremental learning
learn new tasks without forgetting old tasks
- Online learning
training and testing simultaneously
- Batch learning
training followed by testing in seperate activities
- Open Set learning
test samples from classes unseen during training
Approaches To Learning
- Meta-learning
learning to learn more affectively
- Curriculum learning
next training sample selected according to curriculam
- Active learning
next training sample actively rather than stochastically selected
- Domain adaptation
make a model trained on a source domain perform well on a related target domain
- Transfer learning
leverage knowledge gained from one task to improve performance on a related task
- Knowledge distillation
transfer knowledge from a complex model (teacher) to a more efficient model (student)
- Adversarial learning
two models are trained in a competitive setting
- Contrastive learning
learn by contrasting similar and dissimilar samples
- Data augmentation
applying various transformations to a dataset to increase diversity
- Data distillation
create a synthetic dataset that captures the essential knowledge of a larger dataset
Learning Problems
Representation learning
learn a function that maps samples into an embedding space
Metric learning
learn a function that maps samples into an embedded space that preserve the samples’ dis/similarity
Learning-To-Rank
learn a function that maps samples into an embedded space that preserve the samples’ rankings
Detection
select true or false as the task's target
Classification
select one of the classes as the task's target
Recognition
detect (true of false) if sample matches a query target
Retrieval
return all samples that match a query target
Anomaly detection
Out-of-distribution detection
detect instances not from distribution of training - data
Open Set recognition
detect test samples from classes unseen during training
Out-of-domain (OOD) generalization
learn from one or multiple training domains, - to extract a domain-agnostic model which can be applied to an unseen domain
Computer Vision Examples
Object detection
Object tracking
Object recognition
Object re-identification
Instance segmentation
Semantic segmentation
Action/Activity detection
Action/Activity recognition
Depth estimation
Instance Retrieval
Recommendation Systems
Generative models
Density estimation
Bla
Rule-based learning Quantum machine learning Fairness