Data Forest logo
Home page  /  Glossary / 
Semi-supervised Learning

Semi-supervised Learning

Semi-supervised learning is a machine learning approach that combines a small amount of labeled data with a large quantity of unlabeled data during training. This method leverages the limited labeled data to assign structure or guidance, while the larger pool of unlabeled data provides information that aids in improving model accuracy and robustness. Semi-supervised learning is particularly useful when labeled data is scarce or expensive to acquire, and unlabeled data is abundant, a common scenario in real-world applications such as natural language processing, image recognition, and medical diagnostics.

Core Characteristics of Semi-supervised Learning

  1. Data Composition:  
    • In semi-supervised learning, the dataset typically consists of two parts: labeled data (where each instance has an associated label or category) and unlabeled data (where the instances lack such labels).  
    • Labeled data, while more informative, is usually limited in quantity due to the high cost and time required for manual annotation. Unlabeled data, however, is often easier to gather in large amounts, as it does not require labeling.
  2. Learning Paradigm:
    • Semi-supervised learning aims to make use of both labeled and unlabeled data to train a model that generalizes better than one trained solely on labeled data. By extracting patterns or structures from unlabeled data, the model learns to approximate the decision boundaries within the data distribution more effectively.  
    • Typically, semi-supervised learning falls between supervised learning (which relies entirely on labeled data) and unsupervised learning (which relies solely on unlabeled data).
  3. Assumptions and Techniques:
    • luster Assumption: Points within the same cluster are likely to share the same label, so the model assumes that data points close to each other in feature space are of the same class.  
    • Manifold Assumption: Data points lie on a lower-dimensional manifold within the higher-dimensional feature space, and by understanding the manifold structure, the model can infer labels for unlabeled points.  
    • Semi-supervised Learning Techniques:    
    • Self-Training: An iterative process where the model is initially trained on labeled data, then assigns labels to a subset of unlabeled data with high confidence. These pseudo-labeled points are then added to the training set in subsequent iterations.    
    • Co-Training: Two or more classifiers are trained on different feature subsets of the data. Each classifier labels a subset of the unlabeled data, which is used to train the other classifiers.    
    • Graph-based Methods: A graph is constructed from the data, where nodes represent data points, and edges represent similarity or proximity. Labels are propagated through the graph based on the structure, assuming that nearby nodes likely share the same label.

Mathematical Formulation in Semi-supervised Learning

  1. Label Propagation and Graph-Based Formulation:
    • In graph-based semi-supervised learning, the dataset is represented as a graph, with nodes (data points) and weighted edges (similarities). The goal is to propagate labels from labeled nodes to unlabeled nodes by minimizing a loss function based on the edge weights.
    • Let `x_i` represent data points, with `l` labeled points and `u` unlabeled points, and let `f(x_i)` be the label function to be learned. Label propagation involves minimizing:    
      Loss = Σ (f(x_i) - f(x_j))² * W_ij      
      where W_ij represents the weight (similarity) between nodes i and j.
  2. Objective Function in Semi-supervised Loss:
    • Semi-supervised learning combines the supervised loss from labeled data with an unsupervised loss from unlabeled data, which enforces consistency in predictions. For labeled data, the supervised loss is calculated as:    
      L_supervised = (1/n) * Σ L(f(x_i), y_i)      
      where y_i is the true label for labeled data point x_i and L is the loss function (e.g., cross-entropy).  
    • For unlabeled data, the consistency loss encourages similar predictions for similar data points:    
      L_unsupervised = (1/u) * Σ (f(x_i) - f(x_j))²      
      The total loss is then:    
      L_total = L_supervised + λ * L_unsupervised      
      where λ is a regularization parameter that controls the influence of the unlabeled data on model training.
  3. Consistency Regularization:
    Consistency regularization, a common semi-supervised approach, enforces that model predictions are stable under small perturbations or transformations of the input. This means that if an unlabeled data point is modified slightly, the model's output for that point should remain consistent. Formally:    
    Consistency Loss = (1/u) * Σ || f(x_i) - f(T(x_i)) ||²      
    where T(x_i) represents a transformation (e.g., noise addition) applied to x_i.

Semi-supervised learning is especially valuable in fields where acquiring labeled data is costly or infeasible. In medical image analysis, for instance, labeling images requires expert knowledge, making the process slow and resource-intensive. Semi-supervised learning allows these systems to leverage large sets of unlabeled images, supplementing the limited labeled data to improve diagnostic accuracy. Similarly, in natural language processing, semi-supervised learning models can use vast amounts of unlabeled text data to understand language structures better, improving the performance of models used in tasks like sentiment analysis, named entity recognition, and translation.

In Big Data, semi-supervised learning also plays a role in improving scalability. Large-scale data environments often consist of abundant unlabeled data, and using these unlabeled resources helps extend the utility of limited labeled datasets. Semi-supervised learning thus enhances model performance in various real-world scenarios, where resource constraints limit the availability of annotated data.

Data Science
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

Latest publications

All publications
Article preview
January 29, 2025
24 min

AI In Healthcare: Healing by Digital Transformation

Article preview
January 29, 2025
24 min

Predictive Maintenance in Utility Services: Sensor Data for ML

Article preview
January 29, 2025
21 min

Data Science in Power Generation: Energy 4.0 Concept

All publications
top arrow icon