scispace - formally typeset
Open Access

Semi-supervised Learning.

Xiaojin Zhu
- pp 1142-1147
Reads0
Chats0
TLDR
In this article, the authors make a distinction between inductive semi-supervised learning and transductive learning, where the goal is to learn a predictor that predicts future test data better than the predictor learned from the labeled training data alone.
Abstract
Semi-supervised learning uses both labeled and unlabeled data to perform an otherwise supervised learning or unsupervised learning task. In the former case, there is a distinction between inductive semi-supervised learning and transductive learning. In inductive semi-supervised learning, the learner has both labeled training data {(xi, yi)}i=1 iid ∼ p(x, y) and unlabeled training data {xi} i=l+1 iid ∼ p(x), and learns a predictor f : X 7→ Y, f ∈ F where F is the hypothesis space. Here x ∈ X is an input instance, y ∈ Y its target label (discrete for classification or continuous for regression), p(x, y) the unknown joint distribution and p(x) its marginal, and typically l u. The goal is to learn a predictor that predicts future test data better than the predictor learned from the labeled training data alone. In transductive learning, the setting is the same except that one is solely interested in the predictions on the unlabeled training data {xi} i=l+1, without any intention to generalize to future test data. In the latter case, an unsupervised learning task is enhanced by labeled data. For example, in semi-supervised clustering (a.k.a. constrained clustering) one may have a few must-links (two instances must be in the same cluster) and cannot-links (two instances cannot be in the same cluster) in addition to the unlabeled instances to be clustered; in semi-supervised dimensionality reduction one might have the target low-dimensional coordinates on a few instances. This entry will focus on the former case of learning a predictor.

read more

Citations
More filters
Book

Deep Learning

TL;DR: Deep learning as mentioned in this paper is a form of machine learning that enables computers to learn from experience and understand the world in terms of a hierarchy of concepts, and it is used in many applications such as natural language processing, speech recognition, computer vision, online recommendation systems, bioinformatics, and videogames.
Posted Content

Bootstrap Your Own Latent: A New Approach to Self-Supervised Learning

TL;DR: This work introduces Bootstrap Your Own Latent (BYOL), a new approach to self-supervised image representation learning that performs on par or better than the current state of the art on both transfer and semi- supervised benchmarks.
Journal ArticleDOI

A Comprehensive Survey on Transfer Learning

TL;DR: Transfer learning aims to improve the performance of target learners on target domains by transferring the knowledge contained in different but related source domains as discussed by the authors, in which the dependence on a large number of target-domain data can be reduced for constructing target learners.
Posted Content

MixMatch: A Holistic Approach to Semi-Supervised Learning

TL;DR: MixMatch as discussed by the authors predicts low-entropy labels for unlabeled examples and combines them with labeled and unlabelled data using MixUp to obtain state-of-the-art results.
Posted Content

Unsupervised Data Augmentation for Consistency Training

TL;DR: A new perspective on how to effectively noise unlabeled examples is presented and it is argued that the quality of noising, specifically those produced by advanced data augmentation methods, plays a crucial role in semi-supervised learning.
Trending Questions (2)
Whats principal dataset used in semisupervised learning?

The provided paper does not mention a specific principal dataset used in semi-supervised learning.

What are the differences between supervised and semi-supervised learning?

Supervised learning uses labeled data, while semi-supervised learning uses both labeled and unlabeled data.