no code implementations • 29 Jun 2018 • Rhondale Tso, Davide Gerosa, Yanbei Chen
The early inspiral of massive stellar-mass black-hole binaries merging in LIGO's sensitivity band will be detectable at low frequencies by the upcoming space mission LISA.
General Relativity and Quantum Cosmology High Energy Astrophysical Phenomena
1 code implementation • 22 Aug 2018 • Yanbei Chen, Xiatian Zhu, Shaogang Gong
In this work, to address the video person re-id task, we formulate a novel Deep Association Learning (DAL) scheme, the first end-to-end deep learning method using none of the identity labels in model initialisation and training.
Ranked #6 on Person Re-Identification on PRID2011
Unsupervised Person Re-Identification Video-Based Person Re-Identification
1 code implementation • ECCV 2018 • Yanbei Chen, Xiatian Zhu, Shaogang Gong
We consider the semi-supervised multi-class classification problem of learning from sparse labelled and abundant unlabelled training data.
1 code implementation • 31 Jan 2019 • Davide Gerosa, Sizheng Ma, Kaze W. K. Wong, Emanuele Berti, Richard O'Shaughnessy, Yanbei Chen, Krzysztof Belczynski
We use population synthesis simulations of isolated binary stars to explore some of the stellar physics that could be constrained with multiband events, and we show that specific formation pathways might be overrepresented in multiband events compared to ground-only detections.
High Energy Astrophysical Phenomena General Relativity and Quantum Cosmology
no code implementations • ICCV 2019 • Yanbei Chen, Xiatian Zhu, Shaogang Gong
To tackle this limitation, we propose a novel Instance-Guided Context Rendering scheme, which transfers the source person identities into diverse target domain contexts to enable supervised re-id model learning in the unlabelled target domain.
1 code implementation • CVPR 2020 • Yanbei Chen, Shaogang Gong, Loris Bazzani
In this work, we tackle this task by a novel Visiolinguistic Attention Learning (VAL) framework.
Ranked #17 on Image Retrieval on Fashion IQ
1 code implementation • 23 Feb 2021 • Uddeshya Upadhyay, Yanbei Chen, Zeynep Akata
Unpaired image-to-image translation refers to learning inter-image-domain mapping in an unsupervised manner.
1 code implementation • CVPR 2021 • Yanbei Chen, Yongqin Xian, A. Sophia Koepke, Ying Shan, Zeynep Akata
Having access to multi-modal cues (e. g. vision and audio) empowers some cognitive tasks to be done faster compared to learning from a single modality.
no code implementations • 4 May 2021 • Yanbei Chen, Thomas Hummel, A. Sophia Koepke, Zeynep Akata
Recent advances in XAI provide explanations for models trained on still images.
Explainable artificial intelligence Explainable Artificial Intelligence (XAI) +1
1 code implementation • 29 Jun 2021 • Uddeshya Upadhyay, Yanbei Chen, Tobias Hepp, Sergios Gatidis, Zeynep Akata
However, the state-of-the-art GAN-based frameworks do not estimate the uncertainty in the predictions made by the network that is essential for making informed medical decisions and subsequent revision by medical experts and has recently been shown to improve the performance and interpretability of the model.
1 code implementation • NeurIPS 2021 • Uddeshya Upadhyay, Yanbei Chen, Zeynep Akata
Unpaired image-to-image translation refers to learning inter-image-domain mapping without corresponding image pairs.
1 code implementation • 12 Apr 2022 • Andrei Neculai, Yanbei Chen, Zeynep Akata
Without bells and whistles, we show that our probabilistic model formulation significantly outperforms existing related methods on multimodal image retrieval while generalizing well to query with different amounts of inputs given in arbitrary visual and (or) textual modalities.
1 code implementation • 27 Apr 2022 • Ilke Cugu, Massimiliano Mancini, Yanbei Chen, Zeynep Akata
Generalizing visual recognition models trained on a single distribution to unseen input distributions (i. e. domains) requires making them robust to superfluous correlations in the training set.
1 code implementation • 14 Jul 2022 • Uddeshya Upadhyay, Shyamgopal Karthik, Yanbei Chen, Massimiliano Mancini, Zeynep Akata
Moreover, many of the high-performing deep learning models that are already trained and deployed are non-Bayesian in nature and do not provide uncertainty estimates.
no code implementations • 24 Aug 2022 • Yanbei Chen, Massimiliano Mancini, Xiatian Zhu, Zeynep Akata
Semi-supervised learning and unsupervised learning offer promising paradigms to learn from an abundance of unlabeled visual data.
1 code implementation • 19 Oct 2022 • Abhra Chaudhuri, Massimiliano Mancini, Yanbei Chen, Zeynep Akata, Anjan Dutta
Representation learning for sketch-based image retrieval has mostly been tackled by learning embeddings that discard modality-specific information.
1 code implementation • 23 Nov 2022 • Yuchen Ma, Yanbei Chen, Zeynep Akata
In this work, we formulate a new knowledge distillation framework to transfer the knowledge from self-supervised pre-trained models to any other student network by a novel approach named Embedding Graph Alignment.
no code implementations • CVPR 2023 • Yanbei Chen, Manchen Wang, Abhay Mittal, Zhenlin Xu, Paolo Favaro, Joseph Tighe, Davide Modolo
Our results show that ScaleDet achieves compelling strong model performance with an mAP of 50. 7 on LVIS, 58. 8 on COCO, 46. 8 on Objects365, 76. 2 on OpenImages, and 71. 8 on ODinW, surpassing state-of-the-art detectors with the same backbone.
Ranked #1 on Object Detection on OpenImages-v6 (using extra training data)
no code implementations • 28 Jun 2023 • Zhenlin Xu, Yi Zhu, Tiffany Deng, Abhay Mittal, Yanbei Chen, Manchen Wang, Paolo Favaro, Joseph Tighe, Davide Modolo
This paper introduces innovative benchmarks to evaluate Vision-Language Models (VLMs) in real-world zero-shot recognition tasks, focusing on the granularity and specificity of prompting text.
no code implementations • 29 Sep 2023 • Zhuoran Yu, Manchen Wang, Yanbei Chen, Paolo Favaro, Davide Modolo
First, we introduce a denoising scheme to generate reliable pseudo-heatmaps as targets for learning from unlabeled data.
no code implementations • 7 Apr 2024 • Fanjie Kong, Yanbei Chen, Jiarui Cai, Davide Modolo
Specifically, we bootstrap dense synthetic captions using pre-trained VLMs to provide rich descriptions on different regions in images, and incorporate these captions to train a novel detector that generalizes to novel concepts.
1 code implementation • 8 Apr 2024 • Zijia Lu, Bing Shuai, Yanbei Chen, Zhenlin Xu, Davide Modolo
In this paper, we propose a novel concept of path consistency to learn robust object matching without using manual object identity supervision.
no code implementations • ECCV 2020 • Yanbei Chen, Loris Bazzani
Interactive image retrieval is an emerging research topic with the objective of integrating inputs from multiple modalities as query for retrieval, e. g., textual feedback from users to guide, modify or refine image retrieval.