Semi-Supervised Learning
Semi-supervised learningIt is a learning technique between supervised learning and unsupervised learning. It uses both labeled and unlabeled samples for learning.
Semi-supervised learning allows the learner to automatically use unlabeled samples to improve learning performance without relying on external interactions.
Two common semi-supervised learning assumptions
One is the "cluster assumption", which assumes that the data has a cluster structure and samples in the same cluster belong to the same category.
The other type is the “manifold assumption”, which assumes that the data is distributed on a manifold structure and that adjacent samples have similar output values.
Brief introduction to semi-supervised learning theory:
Semi-supervised learning has two sample sets, one labeled and one unlabeled.
Lable = { ( xi , yi ) }, Unlabled = { ( xi ) }. And quantitatively, L << U.
1) Using labeled samples alone, we can generate supervised classification algorithms;
2) Using unlabeled samples alone, we can generate unsupervised clustering algorithms;
3) Using both, we hope to add unlabeled samples in 1 to enhance the effect of supervised classification; similarly, we hope to add labeled samples in 2 to enhance the effect of unsupervised clustering.
Generally speaking, semi-supervised learning focuses on adding unlabeled samples to supervised classification algorithms to achieve semi-supervised classification. That is, unlabeled samples are added to 1 to enhance the classification effect.
Classification of Semi-Supervised Learning Algorithms:
1) Self-training algorithm;
2) Generative models
3) SVMs semi-supervised support vector machines;
4) graph-based methods;
5) Multiview learing algorithm.
References
【1】https://blog.csdn.net/ice110956/article/details/13775071
【2】http://blog.sciencenet.cn/blog-242887-309591.html
【3】https://www.zybuluo.com/Team/note/1133211