Search Results for author: Yanbei Chen

Found 23 papers, 14 papers with code

Optimizing LIGO with LISA forewarnings to improve black-hole spectroscopy

no code implementations29 Jun 2018 Rhondale Tso, Davide Gerosa, Yanbei Chen

The early inspiral of massive stellar-mass black-hole binaries merging in LIGO's sensitivity band will be detectable at low frequencies by the upcoming space mission LISA.

General Relativity and Quantum Cosmology High Energy Astrophysical Phenomena

Deep Association Learning for Unsupervised Video Person Re-identification

1 code implementation22 Aug 2018 Yanbei Chen, Xiatian Zhu, Shaogang Gong

In this work, to address the video person re-id task, we formulate a novel Deep Association Learning (DAL) scheme, the first end-to-end deep learning method using none of the identity labels in model initialisation and training.

Unsupervised Person Re-Identification Video-Based Person Re-Identification

Semi-Supervised Deep Learning with Memory

1 code implementation ECCV 2018 Yanbei Chen, Xiatian Zhu, Shaogang Gong

We consider the semi-supervised multi-class classification problem of learning from sparse labelled and abundant unlabelled training data.

General Classification Multi-class Classification +1

Multiband gravitational-wave event rates and stellar physics

1 code implementation31 Jan 2019 Davide Gerosa, Sizheng Ma, Kaze W. K. Wong, Emanuele Berti, Richard O'Shaughnessy, Yanbei Chen, Krzysztof Belczynski

We use population synthesis simulations of isolated binary stars to explore some of the stellar physics that could be constrained with multiband events, and we show that specific formation pathways might be overrepresented in multiband events compared to ground-only detections.

High Energy Astrophysical Phenomena General Relativity and Quantum Cosmology

Instance-Guided Context Rendering for Cross-Domain Person Re-Identification

no code implementations ICCV 2019 Yanbei Chen, Xiatian Zhu, Shaogang Gong

To tackle this limitation, we propose a novel Instance-Guided Context Rendering scheme, which transfers the source person identities into diverse target domain contexts to enable supervised re-id model learning in the unlabelled target domain.

Generative Adversarial Network Image Generation +1

Uncertainty-aware Generalized Adaptive CycleGAN

1 code implementation23 Feb 2021 Uddeshya Upadhyay, Yanbei Chen, Zeynep Akata

Unpaired image-to-image translation refers to learning inter-image-domain mapping in an unsupervised manner.

Image Denoising Image-to-Image Translation +1

Distilling Audio-Visual Knowledge by Compositional Contrastive Learning

1 code implementation CVPR 2021 Yanbei Chen, Yongqin Xian, A. Sophia Koepke, Ying Shan, Zeynep Akata

Having access to multi-modal cues (e. g. vision and audio) empowers some cognitive tasks to be done faster compared to learning from a single modality.

Audio Tagging audio-visual learning +5

Uncertainty-Guided Progressive GANs for Medical Image Translation

1 code implementation29 Jun 2021 Uddeshya Upadhyay, Yanbei Chen, Tobias Hepp, Sergios Gatidis, Zeynep Akata

However, the state-of-the-art GAN-based frameworks do not estimate the uncertainty in the predictions made by the network that is essential for making informed medical decisions and subsequent revision by medical experts and has recently been shown to improve the performance and interpretability of the model.

Denoising Image-to-Image Translation +2

Robustness via Uncertainty-aware Cycle Consistency

1 code implementation NeurIPS 2021 Uddeshya Upadhyay, Yanbei Chen, Zeynep Akata

Unpaired image-to-image translation refers to learning inter-image-domain mapping without corresponding image pairs.

Autonomous Driving Image-to-Image Translation +1

Probabilistic Compositional Embeddings for Multimodal Image Retrieval

1 code implementation12 Apr 2022 Andrei Neculai, Yanbei Chen, Zeynep Akata

Without bells and whistles, we show that our probabilistic model formulation significantly outperforms existing related methods on multimodal image retrieval while generalizing well to query with different amounts of inputs given in arbitrary visual and (or) textual modalities.

Image Retrieval Retrieval

Attention Consistency on Visual Corruptions for Single-Source Domain Generalization

1 code implementation27 Apr 2022 Ilke Cugu, Massimiliano Mancini, Yanbei Chen, Zeynep Akata

Generalizing visual recognition models trained on a single distribution to unseen input distributions (i. e. domains) requires making them robust to superfluous correlations in the training set.

Domain Generalization

BayesCap: Bayesian Identity Cap for Calibrated Uncertainty in Frozen Neural Networks

1 code implementation14 Jul 2022 Uddeshya Upadhyay, Shyamgopal Karthik, Yanbei Chen, Massimiliano Mancini, Zeynep Akata

Moreover, many of the high-performing deep learning models that are already trained and deployed are non-Bayesian in nature and do not provide uncertainty estimates.

Autonomous Driving Deblurring +2

Semi-Supervised and Unsupervised Deep Visual Learning: A Survey

no code implementations24 Aug 2022 Yanbei Chen, Massimiliano Mancini, Xiatian Zhu, Zeynep Akata

Semi-supervised learning and unsupervised learning offer promising paradigms to learn from an abundance of unlabeled visual data.

Cross-Modal Fusion Distillation for Fine-Grained Sketch-Based Image Retrieval

1 code implementation19 Oct 2022 Abhra Chaudhuri, Massimiliano Mancini, Yanbei Chen, Zeynep Akata, Anjan Dutta

Representation learning for sketch-based image retrieval has mostly been tackled by learning embeddings that discard modality-specific information.

Cross-Modal Retrieval Knowledge Distillation +3

Distilling Knowledge from Self-Supervised Teacher by Embedding Graph Alignment

1 code implementation23 Nov 2022 Yuchen Ma, Yanbei Chen, Zeynep Akata

In this work, we formulate a new knowledge distillation framework to transfer the knowledge from self-supervised pre-trained models to any other student network by a novel approach named Embedding Graph Alignment.

Knowledge Distillation Representation Learning +1

ScaleDet: A Scalable Multi-Dataset Object Detector

no code implementations CVPR 2023 Yanbei Chen, Manchen Wang, Abhay Mittal, Zhenlin Xu, Paolo Favaro, Joseph Tighe, Davide Modolo

Our results show that ScaleDet achieves compelling strong model performance with an mAP of 50. 7 on LVIS, 58. 8 on COCO, 46. 8 on Objects365, 76. 2 on OpenImages, and 71. 8 on ODinW, surpassing state-of-the-art detectors with the same backbone.

 Ranked #1 on Object Detection on OpenImages-v6 (using extra training data)

Object object-detection +1

Benchmarking Zero-Shot Recognition with Vision-Language Models: Challenges on Granularity and Specificity

no code implementations28 Jun 2023 Zhenlin Xu, Yi Zhu, Tiffany Deng, Abhay Mittal, Yanbei Chen, Manchen Wang, Paolo Favaro, Joseph Tighe, Davide Modolo

This paper introduces innovative benchmarks to evaluate Vision-Language Models (VLMs) in real-world zero-shot recognition tasks, focusing on the granularity and specificity of prompting text.

Benchmarking Specificity +1

Denoising and Selecting Pseudo-Heatmaps for Semi-Supervised Human Pose Estimation

no code implementations29 Sep 2023 Zhuoran Yu, Manchen Wang, Yanbei Chen, Paolo Favaro, Davide Modolo

First, we introduce a denoising scheme to generate reliable pseudo-heatmaps as targets for learning from unlabeled data.

Denoising Pose Estimation +1

Hyperbolic Learning with Synthetic Captions for Open-World Detection

no code implementations7 Apr 2024 Fanjie Kong, Yanbei Chen, Jiarui Cai, Davide Modolo

Specifically, we bootstrap dense synthetic captions using pre-trained VLMs to provide rich descriptions on different regions in images, and incorporate these captions to train a novel detector that generalizes to novel concepts.

Hallucination Novel Concepts +3

Self-Supervised Multi-Object Tracking with Path Consistency

1 code implementation8 Apr 2024 Zijia Lu, Bing Shuai, Yanbei Chen, Zhenlin Xu, Davide Modolo

In this paper, we propose a novel concept of path consistency to learn robust object matching without using manual object identity supervision.

Multi-Object Tracking Object

Learning Joint Visual Semantic Matching Embeddings for Language-guided Retrieval

no code implementations ECCV 2020 Yanbei Chen, Loris Bazzani

Interactive image retrieval is an emerging research topic with the objective of integrating inputs from multiple modalities as query for retrieval, e. g., textual feedback from users to guide, modify or refine image retrieval.

Image Retrieval Retrieval +2

Cannot find the paper you are looking for? You can Submit a new open access paper.