Representation Learning

3722 papers with code • 5 benchmarks • 9 datasets

Representation Learning is a process in machine learning where algorithms extract meaningful patterns from raw data to create representations that are easier to understand and process. These representations can be designed for interpretability, reveal hidden features, or be used for transfer learning. They are valuable across many fundamental machine learning tasks like image classification and retrieval.

Deep neural networks can be considered representation learning models that typically encode information which is projected into a different subspace. These representations are then usually passed on to a linear classifier to, for instance, train a classifier.

Representation learning can be divided into:

  • Supervised representation learning: learning representations on task A using annotated data and used to solve task B
  • Unsupervised representation learning: learning representations on a task in an unsupervised way (label-free data). These are then used to address downstream tasks and reducing the need for annotated data when learning news tasks. Powerful models like GPT and BERT leverage unsupervised representation learning to tackle language tasks.

More recently, self-supervised learning (SSL) is one of the main drivers behind unsupervised representation learning in fields like computer vision and NLP.

Here are some additional readings to go deeper on the task:

( Image credit: Visualizing and Understanding Convolutional Networks )

Libraries

Use these libraries to find Representation Learning models and implementations

Latest papers with no code

Deep Representation Learning-Based Dynamic Trajectory Phenotyping for Acute Respiratory Failure in Medical Intensive Care Units

no code yet • 4 May 2024

The study demonstrates the utility of our deep representation learning-based approach in unraveling phenotypes that reflect the heterogeneity in sepsis-induced ARF in terms of different mortality outcomes and severity.

TIPAA-SSL: Text Independent Phone-to-Audio Alignment based on Self-Supervised Learning and Knowledge Transfer

no code yet • 3 May 2024

In this paper, we present a novel approach for text independent phone-to-audio alignment based on phoneme recognition, representation learning and knowledge transfer.

Can We Identify Unknown Audio Recording Environments in Forensic Scenarios?

no code yet • 3 May 2024

However, in forensic investigations, the candidate locations are case-specific.

A Mutual Information Perspective on Federated Contrastive Learning

no code yet • 3 May 2024

Along with the proposed SimCLR extensions, we also study how different sources of non-i. i. d.-ness can impact the performance of federated unsupervised learning through global mutual information maximization; we find that a global objective is beneficial for some sources of non-i. i. d.-ness but can be detrimental for others.

Spatio-Temporal SwinMAE: A Swin Transformer based Multiscale Representation Learner for Temporal Satellite Imagery

no code yet • 3 May 2024

As one of the important application domains of foundation models, earth observation has attracted attention and various approaches have been developed.

The Pyramid of Captions

no code yet • 1 May 2024

Building upon this foundation, we propose a novel Pyramid of Captions (PoCa) method, which constructs caption pyramids by generating localized captions for zoomed-in image patches and integrating them with global caption information using large language models.

Causal Perception Inspired Representation Learning for Trustworthy Image Quality Assessment

no code yet • 30 Apr 2024

In this paper, we propose to build a trustworthy IQA model via Causal Perception inspired Representation Learning (CPRL), and a score reflection attack method for IQA model.

UniFS: Universal Few-shot Instance Perception with Point Representations

no code yet • 30 Apr 2024

In this paper, we propose UniFS, a universal few-shot instance perception model that unifies a wide range of instance perception tasks by reformulating them into a dynamic point representation learning framework.

Understanding Multimodal Contrastive Learning Through Pointwise Mutual Information

no code yet • 30 Apr 2024

Multimodal representation learning to integrate different modalities, such as text, vision, and audio is important for real-world applications.

Principled RLHF from Heterogeneous Feedback via Personalization and Preference Aggregation

no code yet • 30 Apr 2024

We propose two approaches based on reward and preference aggregation, respectively: the former utilizes both utilitarianism and Leximin approaches to aggregate individual reward models, with sample complexity guarantees; the latter directly aggregates the human feedback in the form of probabilistic opinions.