no code implementations • EMNLP 2021 • Yitao Cai, Yue Cao, Xiaojun Wan
Concretely, we transform a sentence into a variety of different semantic or syntactic representations (including AMR, UD, and latent semantic representation), and then decode the sentence back from the semantic representations.
no code implementations • 25 May 2023 • Yihao Huang, Yue Cao, Tianlin Li, Felix Juefei-Xu, Di Lin, Ivor W. Tsang, Yang Liu, Qing Guo
Second, we extend representative adversarial attacks against SAM and study the influence of different prompts on robustness.
no code implementations • 13 Apr 2023 • Yue Cao, Zhen Zhang
By means of the ESO feedback, the plant model is kept as nominal, and hence the structural robustness is achieved for the time-varying internal model.
1 code implementation • 6 Apr 2023 • Xinlong Wang, Xiaosong Zhang, Yue Cao, Wen Wang, Chunhua Shen, Tiejun Huang
We unify various segmentation tasks into a generalist in-context learning framework that accommodates different kinds of segmentation data by transforming them into the same format of images.
Ranked #1 on
Few-Shot Semantic Segmentation
on PASCAL-5i (5-Shot)
1 code implementation • 30 Mar 2023 • Wen Wang, Kangyang Xie, Zide Liu, Hao Chen, Yue Cao, Xinlong Wang, Chunhua Shen
Our vid2vid-zero leverages off-the-shelf image diffusion models, and doesn't require training on any video.
1 code implementation • 27 Mar 2023 • Quan Sun, Yuxin Fang, Ledell Wu, Xinlong Wang, Yue Cao
Our approach incorporates new techniques for representation learning, optimization, and augmentation, enabling EVA-CLIP to achieve superior performance compared to previous CLIP models with the same number of parameters but significantly smaller training costs.
Ranked #4 on
Image Classification
on ObjectNet
(using extra training data)
3 code implementations • 20 Mar 2023 • Yuxin Fang, Quan Sun, Xinggang Wang, Tiejun Huang, Xinlong Wang, Yue Cao
We launch EVA-02, a next-generation Transformer-based visual representation pre-trained to reconstruct strong and robust language-aligned vision features via masked image modeling.
3 code implementations • 12 Mar 2023 • Fan Bao, Shen Nie, Kaiwen Xue, Chongxuan Li, Shi Pu, Yaole Wang, Gang Yue, Yue Cao, Hang Su, Jun Zhu
Inspired by the unified view, UniDiffuser learns all distributions simultaneously with a minimal modification to the original diffusion model -- perturbs data in all modalities instead of a single modality, inputs individual timesteps in different modalities, and predicts the noise of all modalities instead of a single modality.
no code implementations • 24 Feb 2023 • Yue Cao, C. S. George Lee
To cope with this issue, we propose a novel behavior-tree-based task generation approach that utilizes state-of-the-art large language models.
1 code implementation • 5 Feb 2023 • Chenyu Zheng, Guoqiang Wu, Fan Bao, Yue Cao, Chongxuan Li, Jun Zhu
Theoretically, the paper considers the surrogate loss instead of the zero-one loss in analyses and generalizes the classical results from binary cases to multiclass ones.
no code implementations • CVPR 2023 • Yixuan Wei, Yue Cao, Zheng Zhang, Houwen Peng, Zhuliang Yao, Zhenda Xie, Han Hu, Baining Guo
This paper presents a method that effectively combines two prevalent visual recognition methods, i. e., image classification and contrastive language-image pre-training, dubbed iCLIP.
no code implementations • CVPR 2023 • Yue Cao, Ming Liu, Shuai Liu, Xiaotao Wang, Lei Lei, WangMeng Zuo
Although deep neural networks have achieved astonishing performance in many vision tasks, existing learning-based methods are far inferior to the physical model-based solutions in extreme low-light sensor noise modeling.
3 code implementations • 8 Dec 2022 • Zanlin Ni, Yulin Wang, Jiangwei Yu, Haojun Jiang, Yue Cao, Gao Huang
In this paper, we present Deep Incubation, a novel approach that enables the efficient and effective training of large models by dividing them into smaller sub-modules that can be trained separately and assembled seamlessly.
1 code implementation • CVPR 2023 • Xinlong Wang, Wen Wang, Yue Cao, Chunhua Shen, Tiejun Huang
In this work, we present Painter, a generalist model which addresses these obstacles with an "image"-centric solution, that is, to redefine the output of core vision tasks as images, and specify task prompts as also images.
Ranked #6 on
Personalized Segmentation
on PerSeg
5 code implementations • CVPR 2023 • Yuxin Fang, Wen Wang, Binhui Xie, Quan Sun, Ledell Wu, Xinggang Wang, Tiejun Huang, Xinlong Wang, Yue Cao
We launch EVA, a vision-centric foundation model to explore the limits of visual representation at scale using only publicly accessible data.
no code implementations • 3 Nov 2022 • Yutong Lin, Ze Liu, Zheng Zhang, Han Hu, Nanning Zheng, Stephen Lin, Yue Cao
In this paper, we present a study of frozen pretrained models when applied to diverse and representative computer vision tasks, including object detection, semantic segmentation and video action recognition.
Ranked #3 on
Action Recognition In Videos
on Kinetics-400
2 code implementations • CVPR 2023 • Fan Bao, Shen Nie, Kaiwen Xue, Yue Cao, Chongxuan Li, Hang Su, Jun Zhu
We evaluate U-ViT in unconditional and class-conditional image generation, as well as text-to-image generation tasks, where U-ViT is comparable if not superior to a CNN-based U-Net of a similar size.
no code implementations • 2 Sep 2022 • Yue Cao, Shaoshi Yang, Zhiyong Feng
We propose a geographic and spatio-temporal information based distributed cooperative positioning (GSTICP) algorithm for wireless networks that require three-dimensional (3D) coordinates and operate in the line-of-sight (LOS) and nonline-of-sight (NLOS) mixed environments.
no code implementations • 25 Aug 2022 • Yue Cao, Shaoshi Yang, Zhiyong Feng, Lihua Wang, Lajos Hanzo
A distributed spatio-temporal information based cooperative positioning (STICP) algorithm is proposed for wireless networks that require three-dimensional (3D) coordinates and operate in the global navigation satellite system (GNSS) denied environments.
no code implementations • 7 Jul 2022 • Yue Cao, Xiaojiang Zhou, Peihao Huang, Yao Xiao, Dayao Chen, Sheng Chen
In this paper, we focus on the information transfer from ranking to pre-ranking stage.
1 code implementation • CVPR 2023 • Zhenda Xie, Zheng Zhang, Yue Cao, Yutong Lin, Yixuan Wei, Qi Dai, Han Hu
Our study reveals that: (i) Masked image modeling is also demanding on larger data.
1 code implementation • 27 May 2022 • Yixuan Wei, Han Hu, Zhenda Xie, Zheng Zhang, Yue Cao, Jianmin Bao, Dong Chen, Baining Guo
These properties, which we aggregately refer to as optimization friendliness, are identified and analyzed by a set of attention- and optimization-related diagnosis tools.
Ranked #3 on
Instance Segmentation
on COCO test-dev
(using extra training data)
1 code implementation • CVPR 2023 • Zhenda Xie, Zigang Geng, Jingcheng Hu, Zheng Zhang, Han Hu, Yue Cao
In this paper, we compare MIM with the long-dominant supervised pre-trained models from two perspectives, the visualizations and the experiments, to uncover their key representational differences.
Ranked #1 on
Monocular Depth Estimation
on KITTI Eigen split
1 code implementation • 20 May 2022 • Yue Cao, Xiaojiang Zhou, Jiaqi Feng, Peihao Huang, Yao Xiao, Dayao Chen, Sheng Chen
However, the retrieval-based methods are sub-optimal and would cause more or less information losses, and it's difficult to balance the effectiveness and efficiency of the retrieval algorithm.
no code implementations • 22 Apr 2022 • Yixuan Wei, Yue Cao, Zheng Zhang, Zhuliang Yao, Zhenda Xie, Han Hu, Baining Guo
Second, we convert the image classification problem from learning parametric category classifier weights to learning a text encoder as a meta network to generate category classifier weights.
no code implementations • CVPR 2022 • Yue Cao, Zhaolin Wan, Dongwei Ren, Zifei Yan, WangMeng Zuo
Particularly, by treating all labeled data as positive samples, PU learning is leveraged to identify negative samples (i. e., outliers) from unlabeled data.
no code implementations • 23 Mar 2022 • Yue Cao, Zhen Zhang
Among the existing contouring control methods, the cross coupled control (CCC) lacks of an asymptotical tracking performance for general contours, and the task coordinate frame (TCF) control usually leads to system nonlinearity, and by design is not well-suited for multi-axis contour tracking.
no code implementations • Global Change Biology 2022 • Huiwen Li, Yiping Wu, Shuguang Liu, Jingfeng Xiao, Wenzhi Zhao, Ji Chen, Georgii Alexandrov, Yue Cao
Additionally, the national cropland topsoil organic carbon increased with a rate of 23. 6 ± 7. 6 g C m−2 yr−1since the 1980s, and the widely applied nitrogenous fertilizer was a key stimulus.
no code implementations • CVPR 2022 • Fei Xie, Chunyu Wang, Guangting Wang, Yue Cao, Wankou Yang, Wenjun Zeng
In contrast to the Siamese-like feature extraction, our network deeply embeds cross-image feature correlation in multiple layers of the feature network.
no code implementations • 4 Feb 2022 • Yue Cao, Fatemeh H. Fard
In this paper, we evaluate PTMs to generate replies to the mobile app user feedbacks.
no code implementations • 4 Jan 2022 • Florin C. Ghesu, Bogdan Georgescu, Awais Mansoor, Youngjin Yoo, Dominik Neumann, Pragneshkumar Patel, R. S. Vishwanath, James M. Balter, Yue Cao, Sasa Grbic, Dorin Comaniciu
Building accurate and robust artificial intelligence systems for medical image assessment requires not only the research and design of advanced deep learning models but also the creation of large and curated sets of annotated training examples.
1 code implementation • 29 Dec 2021 • Mengde Xu, Zheng Zhang, Fangyun Wei, Yutong Lin, Yue Cao, Han Hu, Xiang Bai
However, semantic segmentation and the CLIP model perform on different visual granularity, that semantic segmentation processes on pixels while CLIP performs on images.
Ranked #1 on
Open Vocabulary Semantic Segmentation
on Cityscapes
16 code implementations • CVPR 2022 • Ze Liu, Han Hu, Yutong Lin, Zhuliang Yao, Zhenda Xie, Yixuan Wei, Jia Ning, Yue Cao, Zheng Zhang, Li Dong, Furu Wei, Baining Guo
Three main techniques are proposed: 1) a residual-post-norm method combined with cosine attention to improve training stability; 2) A log-spaced continuous position bias method to effectively transfer models pre-trained using low-resolution images to downstream tasks with high-resolution inputs; 3) A self-supervised pre-training method, SimMIM, to reduce the needs of vast labeled images.
Ranked #4 on
Image Classification
on ImageNet V2
(using extra training data)
2 code implementations • CVPR 2022 • Zhenda Xie, Zheng Zhang, Yue Cao, Yutong Lin, Jianmin Bao, Zhuliang Yao, Qi Dai, Han Hu
We also leverage this approach to facilitate the training of a 3B model (SwinV2-G), that by $40\times$ less data than that in previous practice, we achieve the state-of-the-art on four representative vision benchmarks.
Representation Learning
Self-Supervised Image Classification
1 code implementation • NeurIPS 2021 • Mengde Xu, Zheng Zhang, Fangyun Wei, Yutong Lin, Yue Cao, Stephen Lin, Han Hu, Xiang Bai
We introduce MixTraining, a new training paradigm for object detection that can improve the performance of existing detectors for free.
no code implementations • ICLR 2022 • Yuning You, Yue Cao, Tianlong Chen, Zhangyang Wang, Yang shen
Optimizing an objective function with uncertainty awareness is well-known to improve the accuracy and confidence of optimization solutions.
no code implementations • 23 Sep 2021 • Muhammad Khalid, Liang Wang, Kezhi Wang, Cunhua Pan, Nauman Aslam, Yue Cao
In this paper, to reduce the congestion rate at the city center and increase the quality of experience (QoE) of each user, the framework of long-range autonomous valet parking (LAVP) is presented, where an Autonomous Vehicle (AV) is deployed in the city, which can pick up, drop off users at their required spots, and then drive to the car park out of city center autonomously.
1 code implementation • 24 Jun 2021 • Yue Cao, Payel Das, Vijil Chenthamarakshan, Pin-Yu Chen, Igor Melnyk, Yang shen
Designing novel protein sequences for a desired 3D topological fold is a fundamental yet non-trivial task in protein engineering.
13 code implementations • CVPR 2022 • Ze Liu, Jia Ning, Yue Cao, Yixuan Wei, Zheng Zhang, Stephen Lin, Han Hu
The vision community is witnessing a modeling shift from CNNs to Transformers, where pure Transformer architectures have attained top accuracy on the major video recognition benchmarks.
Ranked #25 on
Action Classification
on Kinetics-600
(using extra training data)
no code implementations • NAACL 2021 • Yue Cao, Hao-Ran Wei, Boxing Chen, Xiaojun Wan
In practical applications, NMT models are usually trained on a general domain corpus and then fine-tuned by continuing training on the in-domain corpus.
1 code implementation • 12 May 2021 • Yansong Tang, Zhenyu Jiang, Zhenda Xie, Yue Cao, Zheng Zhang, Philip H. S. Torr, Han Hu
Previous cycle-consistency correspondence learning methods usually leverage image patches for training.
3 code implementations • 10 May 2021 • Zhenda Xie, Yutong Lin, Zhuliang Yao, Zheng Zhang, Qi Dai, Yue Cao, Han Hu
We are witnessing a modeling shift from CNN to Transformers in computer vision.
Ranked #62 on
Self-Supervised Image Classification
on ImageNet
no code implementations • Archives of Medical Science 2021 • Cheng Xu, Jing Wang, TianLong Zheng, Yue Cao, Fan Ye
Among the 10 public datasets, the Random Forest weighted in accuracy has the best performance on 6 datasets, with an average increase of 1. 44% in accuracy and an average increase of 1. 2% in AUC.
3 code implementations • ICCV 2021 • Ze Liu, Zheng Zhang, Yue Cao, Han Hu, Xin Tong
Instead of grouping local points to each object candidate, our method computes the feature of an object from all the points in the point cloud with the help of an attention mechanism in the Transformers \cite{vaswani2017attention}, where the contribution of each point is automatically learned in the network training.
Ranked #3 on
3D Object Detection
on SUN-RGBD
66 code implementations • ICCV 2021 • Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, Baining Guo
This paper presents a new vision Transformer, called Swin Transformer, that capably serves as a general-purpose backbone for computer vision.
Ranked #2 on
Image Classification
on OmniBenchmark
1 code implementation • 18 Mar 2021 • Yue Cao, Xiaohe Wu, Shuran Qi, Xiao Liu, Zhongqin Wu, WangMeng Zuo
To begin with, the pre-trained denoiser is used to generate the pseudo clean images for the test images.
1 code implementation • EACL 2021 • Qingxiu Dong, Xiaojun Wan, Yue Cao
We propose ParaSCI, the first large-scale paraphrase dataset in the scientific field, including 33, 981 paraphrase pairs from ACL (ParaSCI-ACL) and 316, 063 pairs from arXiv (ParaSCI-arXiv).
no code implementations • 1 Jan 2021 • Yue Cao, Tianlong Chen, Zhangyang Wang, Yang shen
Optimizing an objective function with uncertainty awareness is well-known to improve the accuracy and confidence of optimization solutions.
1 code implementation • ICCVW 2021 • Zhuliang Yao, Yue Cao, Yutong Lin, Ze Liu, Zheng Zhang, Han Hu
Transformer-based vision architectures have attracted great attention because of the strong performance over the convolutional neural networks (CNNs).
3 code implementations • 24 Dec 2020 • Yue Cao, Jiarui Xu, Stephen Lin, Fangyun Wei, Han Hu
The Non-Local Network (NLNet) presents a pioneering approach for capturing long-range dependencies within an image, via aggregating query-specific global context to each query position.
Ranked #37 on
Instance Segmentation
on COCO test-dev
no code implementations • 4 Dec 2020 • Bo Hu, Keping Qiu, Yue Cao, Junhao Liu, Yuwei Wang, Guangxing Li, Zhiqiang Shen, Juan Li, Junzhi Wang, Bin Li, Jian Dong
DR21 south filament (DR21SF) is a unique component of the giant network of filamentary molecular clouds in the north region of Cygnus X complex.
Astrophysics of Galaxies
7 code implementations • CVPR 2021 • Zhenda Xie, Yutong Lin, Zheng Zhang, Yue Cao, Stephen Lin, Han Hu
We argue that the power of contrastive learning has yet to be fully unleashed, as current methods are trained only on instance-level pretext tasks, leading to representations that may be sub-optimal for downstream tasks requiring dense pixel predictions.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Yue Cao, Xiaojun Wan
In this paper, we propose a deep generative model to generate diverse paraphrases.
2 code implementations • 23 Oct 2020 • Yali Peng, Yue Cao, Shigang Liu, Jian Yang, WangMeng Zuo
To cope with this issue, this paper presents a multi-level wavelet residual network (MWRN) architecture as well as a progressive training (PTMWRN) scheme to improve image denoising performance.
2 code implementations • ECCV 2020 • Xiaohe Wu, Ming Liu, Yue Cao, Dongwei Ren, WangMeng Zuo
As for knowledge distillation, we first apply the learned noise models to clean images to synthesize a paired set of training images, and use the real noisy images and the corresponding denoising results in the first stage to form another paired set.
1 code implementation • NeurIPS 2020 • Yihong Chen, Zheng Zhang, Yue Cao, Li-Wei Wang, Stephen Lin, Han Hu
Though RepPoints provides high performance, we find that its heavy reliance on regression for object localization leaves room for improvement.
Ranked #70 on
Object Detection
on COCO test-dev
no code implementations • 8 Jul 2020 • Florin C. Ghesu, Bogdan Georgescu, Awais Mansoor, Youngjin Yoo, Eli Gibson, R. S. Vishwanath, Abishek Balachandran, James M. Balter, Yue Cao, Ramandeep Singh, Subba R. Digumarthy, Mannudeep K. Kalra, Sasa Grbic, Dorin Comaniciu
In our experiments we demonstrate that sample rejection based on the predicted uncertainty can significantly improve the ROC-AUC for various tasks, e. g., by 8% to 0. 91 with an expected rejection rate of under 25% for the classification of different abnormalities in chest radiographs.
1 code implementation • ECCV 2020 • Ze Liu, Han Hu, Yue Cao, Zheng Zhang, Xin Tong
Our investigation reveals that despite the different designs of these operators, all of these operators make surprisingly similar contributions to the network performance under the same network input and feature numbers and result in the state-of-the-art accuracy on standard benchmarks.
Ranked #4 on
3D Semantic Segmentation
on PartNet
no code implementations • ACL 2020 • Yue Cao, Hui Liu, Xiaojun Wan
However, it is a big challenge for the model to directly learn cross-lingual summarization as it requires learning to understand different languages and learning how to summarize at the same time.
1 code implementation • NeurIPS 2020 • Yue Cao, Zhenda Xie, Bin Liu, Yutong Lin, Zheng Zhang, Han Hu
This paper presents parametric instance classification (PIC) for unsupervised visual feature learning.
4 code implementations • ECCV 2020 • Minghao Yin, Zhuliang Yao, Yue Cao, Xiu Li, Zheng Zhang, Stephen Lin, Han Hu
This paper first studies the non-local block in depth, where we find that its attention computation can be split into two terms, a whitened pairwise term accounting for the relationship between two pixels and a unary term representing the saliency of every pixel.
Ranked #14 on
Semantic Segmentation
on Cityscapes test
1 code implementation • 8 May 2020 • Abdelrahman Abdelhamed, Mahmoud Afifi, Radu Timofte, Michael S. Brown, Yue Cao, Zhilu Zhang, WangMeng Zuo, Xiaoling Zhang, Jiye Liu, Wendong Chen, Changyuan Wen, Meng Liu, Shuailin Lv, Yunchao Zhang, Zhihong Pan, Baopu Li, Teng Xi, Yanwen Fan, Xiyu Yu, Gang Zhang, Jingtuo Liu, Junyu Han, Errui Ding, Songhyun Yu, Bumjun Park, Jechang Jeong, Shuai Liu, Ziyao Zong, Nan Nan, Chenghua Li, Zengli Yang, Long Bao, Shuangquan Wang, Dongwoon Bai, Jungwon Lee, Youngjung Kim, Kyeongha Rho, Changyeop Shin, Sungho Kim, Pengliang Tang, Yiyun Zhao, Yuqian Zhou, Yuchen Fan, Thomas Huang, Zhihao LI, Nisarg A. Shah, Wei Liu, Qiong Yan, Yuzhi Zhao, Marcin Możejko, Tomasz Latkowski, Lukasz Treszczotko, Michał Szafraniuk, Krzysztof Trojanowski, Yanhong Wu, Pablo Navarrete Michelini, Fengshuo Hu, Yunhua Lu, Sujin Kim, Wonjin Kim, Jaayeon Lee, Jang-Hwan Choi, Magauiya Zhussip, Azamat Khassenov, Jong Hyun Kim, Hwechul Cho, Priya Kansal, Sabari Nathan, Zhangyu Ye, Xiwen Lu, Yaqi Wu, Jiangxin Yang, Yanlong Cao, Siliang Tang, Yanpeng Cao, Matteo Maggioni, Ioannis Marras, Thomas Tanay, Gregory Slabaugh, Youliang Yan, Myungjoo Kang, Han-Soo Choi, Kyungmin Song, Shusong Xu, Xiaomu Lu, Tingniao Wang, Chunxia Lei, Bin Liu, Rajat Gupta, Vineet Kumar
This challenge is based on a newly collected validation and testing image datasets, and hence, named SIDD+.
2 code implementations • CVPR 2020 • Yihong Chen, Yue Cao, Han Hu, Li-Wei Wang
We argue that there are two important cues for humans to recognize objects in videos: the global semantic information and the local localization information.
Ranked #7 on
Video Object Detection
on ImageNet VID
1 code implementation • ECCV 2020 • Bin Liu, Yue Cao, Yutong Lin, Qi Li, Zheng Zhang, Mingsheng Long, Han Hu
This paper introduces a negative margin loss to metric learning based few-shot learning methods.
2 code implementations • CVPR 2021 • Zhuliang Yao, Yue Cao, Shuxin Zheng, Gao Huang, Stephen Lin
We thus compensate for the network weight changes via a proposed technique based on Taylor polynomials, so that the statistics can be accurately estimated and batch normalization can be effectively applied.
Ranked #193 on
Object Detection
on COCO test-dev
no code implementations • 28 Dec 2019 • Yue Cao, Yang shen
Moreover, estimating model quality, also known as the quality assessment problem, is rarely addressed in protein docking.
1 code implementation • NeurIPS 2019 • Yue Cao, Tianlong Chen, Zhangyang Wang, Yang shen
Learning to optimize has emerged as a powerful framework for various optimization and machine learning tasks.
3 code implementations • ICLR 2020 • Weijie Su, Xizhou Zhu, Yue Cao, Bin Li, Lewei Lu, Furu Wei, Jifeng Dai
We introduce a new pre-trainable generic representation for visual-linguistic tasks, called Visual-Linguistic BERT (VL-BERT for short).
Ranked #1 on
Visual Question Answering (VQA)
on VCR (Q-A) dev
9 code implementations • 25 Apr 2019 • Yue Cao, Jiarui Xu, Stephen Lin, Fangyun Wei, Han Hu
In this paper, we take advantage of this finding to create a simplified network based on a query-independent formulation, which maintains the accuracy of NLNet but with significantly less computation.
Ranked #55 on
Instance Segmentation
on COCO test-dev
no code implementations • ICCV 2019 • Jiarui Xu, Yue Cao, Zheng Zhang, Han Hu
Recent progress in multiple object tracking (MOT) has shown that a robust similarity score is key to the success of trackers.
1 code implementation • 9 Mar 2019 • Weizhi Ma, Min Zhang, Yue Cao, Woojeong, Jin, Chenyang Wang, Yiqun Liu, Shaoping Ma, Xiang Ren
The framework encourages two modules to complement each other in generating effective and explainable recommendation: 1) inductive rules, mined from item-centric knowledge graphs, summarize common multi-hop relational patterns for inferring different item associations and provide human-readable explanation for model prediction; 2) recommendation module can be augmented by induced rules and thus have better generalization ability dealing with the cold-start issue.
1 code implementation • 1 Feb 2019 • Bin Liu, Yue Cao, Mingsheng Long, Jian-Min Wang, Jingdong Wang
We propose Deep Triplet Quantization (DTQ), a novel approach to learning deep quantization models from the similarity triplets.
Ranked #1 on
Image Retrieval
on NUS-WIDE
1 code implementation • 31 Jan 2019 • Yue Cao, Yang shen
To the best of our knowledge, this study represents the first uncertainty quantification solution for protein docking, with theoretical rigor and comprehensive assessment.
no code implementations • CVPR 2018 • Yue Cao, Bin Liu, Mingsheng Long, Jian-Min Wang
The main idea is to augment the training data with nearly real images synthesized from a new Pair Conditional Wasserstein GAN (PC-WGAN) conditioned on the pairwise similarity information.
no code implementations • CVPR 2018 • Yue Cao, Mingsheng Long, Bin Liu, Jian-Min Wang
Due to its computation efficiency and retrieval quality, hashing has been widely applied to approximate nearest neighbor search for large-scale image retrieval, while deep hashing further improves the retrieval quality by end-to-end representation learning and hash coding.
no code implementations • CVPR 2017 • Yue Cao, Mingsheng Long, Jian-Min Wang, Shichen Liu
This paper presents a compact coding solution with a focus on the deep learning to quantization approach, which improves retrieval quality by end-to-end representation learning and compact encoding and has already shown the superior performance over the hashing solutions for similarity retrieval.
no code implementations • 22 Feb 2016 • Yue Cao, Mingsheng Long, Jian-Min Wang, Philip S. Yu
This paper presents a Correlation Hashing Network (CHN) approach to cross-modal hashing, which jointly learns good data representation tailored to hash coding and formally controls the quantization error.
5 code implementations • 10 Feb 2015 • Mingsheng Long, Yue Cao, Jian-Min Wang, Michael. I. Jordan
Recent studies reveal that a deep neural network can learn transferable features which generalize well to novel tasks for domain adaptation.
Ranked #3 on
Domain Adaptation
on Synth Digits-to-SVHN