1 code implementation • ECCV 2020 • Guangrui Li, Guoliang Kang, Wu Liu, Yunchao Wei, Yi Yang
The target of CCM is to acquire those synthetic images that share similar distribution with the real ones in the target domain, so that the domain gap can be naturally alleviated by employing the content-consistent synthetic images for training.
Ranked #12 on Semantic Segmentation on GTAV-to-Cityscapes Labels
no code implementations • 2 Dec 2024 • Jiakai Wang, Pengfei Zhang, Renshuai Tao, Jian Yang, Hao liu, Xianglong Liu, Yunchao Wei, Yao Zhao
Specifically, to adapt the optimization goal of behavior backdoor, we introduce the behavior-driven backdoor object optimizing method by a bi-target behavior backdoor training loss, thus we could guide the poisoned model optimization direction.
no code implementations • 30 Nov 2024 • Weizhe Liu, Renshuai Tao, Hongguang Zhu, YunDa Sun, Yao Zhao, Yunchao Wei
The approach introduces 1) contour information of baggage and 2) variation of material information into the original image by Mixup at patch level.
no code implementations • 29 Nov 2024 • Kunyang Han, Yibo Hu, Mengxue Qu, Hailin Shi, Yao Zhao, Yunchao Wei
Advances in CLIP and large multimodal models (LMMs) have enabled open-vocabulary and free-text segmentation, yet existing models still require predefined category prompts, limiting free-form category self-generation.
no code implementations • 27 Nov 2024 • Renshuai Tao, Haoyu Wang, Wei Wang, Yunchao Wei, Yao Zhao
The detection of prohibited items in X-ray security inspections is vital for ensuring public safety.
no code implementations • 27 Nov 2024 • Renshuai Tao, Haoyu Wang, Yuzhe Guo, Hairong Chen, Li Zhang, Xianglong Liu, Yunchao Wei, Yao Zhao
To emulate human intelligence in dual-view detection, we propose the Auxiliary-view Enhanced Network (AENet), a novel detection framework that leverages both the main and auxiliary views of the same object.
1 code implementation • 22 Nov 2024 • Wanqi Yang, Yanda Li, Meng Fang, Yunchao Wei, Tianyi Zhou, Ling Chen
We evaluate six state-of-the-art LLMs with voice interaction capabilities, including Gemini-1. 5-Pro, GPT-4o, and others, using three distinct evaluation methods on the CAA benchmark.
no code implementations • 20 Nov 2024 • Xinhao Zhong, Siyu Jiao, Yao Zhao, Yunchao Wei
However, in open-set scenarios, the unlabeled dataset contains both in-distribution (ID) classes and out-of-distribution (OOD) classes.
1 code implementation • 4 Nov 2024 • Biao Wu, Yanda Li, Meng Fang, Zirui Song, Zhiwei Zhang, Yunchao Wei, Ling Chen
This survey provides a comprehensive review of mobile agent technologies, focusing on recent advancements that enhance real-time adaptability and multimodal interaction.
no code implementations • 15 Oct 2024 • Man Liu, Huihui Bai, Feng Li, Chunjie Zhang, Yunchao Wei, Meng Wang, Tat-Seng Chua, Yao Zhao
Generalized zero-shot learning (GZSL) endeavors to identify the unseen categories using knowledge from the seen domain, necessitating the intrinsic interactions between the visual features and attribute semantic features.
1 code implementation • 9 Oct 2024 • Anqi Zhang, Guangyu Gao, Jianbo Jiao, Chi Harold Liu, Yunchao Wei
Another subsequent Point-Mask Clustering module aligns the granularity of masks and selected points as a directed graph, based on mask coverage over points.
Ranked #2 on Few-Shot Semantic Segmentation on COCO-20i (5-shot)
1 code implementation • 6 Oct 2024 • Jingxuan Xu, Wuyang Chen, Linyi Li, Yao Zhao, Yunchao Wei
To mitigate societal biases implicitly encoded in recent successful pretrained language models, a diverse array of approaches have been proposed to encourage model fairness, focusing on prompting, data augmentation, regularized fine-tuning, and more.
1 code implementation • 15 Aug 2024 • Gengwei Zhang, Liyuan Wang, Guoliang Kang, Ling Chen, Yunchao Wei
Considering that the overly fast representation learning and the biased classification layer constitute this particular problem, we introduce the advanced Slow Learner with Classifier Alignment (SLCA++) framework to unleash the power of Seq FT, serving as a strong baseline approach for CLPT.
1 code implementation • 6 Aug 2024 • Yiming Zhong, Xiaolin Zhang, Yao Zhao, Yunchao Wei
Secondly, we propose a dual timestep strategy, increasing the consistency of guidance and optimizing 3D models from geometry to appearance in DreamLCM.
no code implementations • 5 Aug 2024 • Yanda Li, Chi Zhang, Wanqi Yang, Bin Fu, Pei Cheng, Xin Chen, Ling Chen, Yunchao Wei
In the deployment phase, RAG technology enables efficient retrieval and update from this knowledge base, thereby empowering the agent to perform tasks effectively and accurately.
1 code implementation • 1 Aug 2024 • Siyu Jiao, Hongguang Zhu, Jiannan Huang, Yao Zhao, Yunchao Wei, Humphrey Shi
In this way, the vision and text representation of CLIP are optimized collaboratively, enhancing the alignment of the vision-text feature space.
Ranked #2 on Open Vocabulary Semantic Segmentation on ADE20K-847
Open Vocabulary Panoptic Segmentation Open Vocabulary Semantic Segmentation
no code implementations • 3 Jul 2024 • Weitai Kang, Mengxue Qu, Yunchao Wei, Yan Yan
Building upon this, ACTRESS consists of an active sampling strategy and a selective retraining strategy.
2 code implementations • 24 Jun 2024 • Henghui Ding, Chang Liu, Yunchao Wei, Nikhila Ravi, Shuting He, Song Bai, Philip Torr, Deshui Miao, Xin Li, Zhenyu He, YaoWei Wang, Ming-Hsuan Yang, Zhensong Xu, Jiangtao Yao, Chengjing Wu, Ting Liu, Luoqi Liu, Xinyu Liu, Jing Zhang, Kexin Zhang, Yuting Yang, Licheng Jiao, Shuyuan Yang, Mingqi Gao, Jingnan Luo, Jinyu Yang, Jungong Han, Feng Zheng, Bin Cao, Yisi Zhang, Xuanxu Lin, Xingjian He, Bo Zhao, Jing Liu, Feiyu Pan, Hao Fang, Xiankai Lu
Moreover, we provide a new motion expression guided video segmentation dataset MeViS to study the natural language-guided video understanding in complex environments.
no code implementations • 20 Jun 2024 • Chaoqi Liang, Guanglei Yang, Lifeng Qiao, Zitong Huang, Hongliang Yan, Yunchao Wei, WangMeng Zuo
Our approach, LayerMatch, which integrates these two strategies, can avoid the severe interference of noisy pseudo-labels in the linear classification layer while accelerating the clustering capability of the feature extraction layer.
1 code implementation • CVPR 2024 • Bingfeng Zhang, Siyue Yu, Yunchao Wei, Yao Zhao, Jimin Xiao
Specifically, the frozen CLIP model is applied as the backbone for semantic feature extraction, and a new decoder is designed to interpret extracted semantic features for final prediction.
no code implementations • 5 Jun 2024 • Man Liu, Huihui Bai, Feng Li, Chunjie Zhang, Yunchao Wei, Meng Wang, Tat-Seng Chua, Yao Zhao
Zero-shot learning (ZSL) aims to explore the semantic-visual interactions to discover comprehensive knowledge transferred from seen categories to classify unseen categories.
no code implementations • 28 May 2024 • Weitai Kang, Mengxue Qu, Jyoti Kini, Yunchao Wei, Mubarak Shah, Yan Yan
To achieve detection based on human intention, it relies on humans to observe the scene, reason out the target that aligns with their intention ("pillow" in this case), and finally provide a reference to the AI system, such as "A pillow on the couch".
1 code implementation • 27 May 2024 • Jiannan Huang, Jun Hao Liew, Hanshu Yan, Yuyang Yin, Yao Zhao, Yunchao Wei
Recent text-to-image customization works have been proven successful in generating images of given concepts by fine-tuning the diffusion models on a few examples.
no code implementations • 27 May 2024 • Jian Zhao, Lei Jin, Jianshu Li, Zheng Zhu, Yinglei Teng, Jiaojiao Zhao, Sadaf Gulshad, Zheng Wang, Bo Zhao, Xiangbo Shu, Yunchao Wei, Xuecheng Nie, Xiaojie Jin, Xiaodan Liang, Shin'ichi Satoh, Yandong Guo, Cewu Lu, Junliang Xing, Jane Shen Shengmei
The SkatingVerse Workshop & Challenge aims to encourage research in developing novel and accurate methods for human action understanding.
no code implementations • 26 May 2024 • Hanwen Liang, Yuyang Yin, Dejia Xu, Hanxue Liang, Zhangyang Wang, Konstantinos N. Plataniotis, Yao Zhao, Yunchao Wei
Building on this foundation, we propose a strategy to migrate the temporal consistency in video diffusion models to the spatial-temporal consistency required for 4D generation.
1 code implementation • CVPR 2024 • Jingxuan Xu, Wuyang Chen, Yao Zhao, Yunchao Wei
In the context of efficient OVS, we target achieving performance that is comparable to or even better than prior OVS works based on large vision-language foundation models, by utilizing smaller models that incur lower training costs.
1 code implementation • 30 Mar 2024 • Chenyi Zhang, Yihan Hu, Henghui Ding, Humphrey Shi, Yao Zhao, Yunchao Wei
Despite significant advancements in image matting, existing models heavily depend on manually-drawn trimaps for accurate results in natural image scenarios.
1 code implementation • 20 Mar 2024 • Linshan Wu, Zhun Zhong, Jiayi Ma, Yunchao Wei, Hao Chen, Leyuan Fang, Shutao Li
Based on the label distributions, we leverage the GMM to generate high-quality pseudo labels for more reliable supervision.
Weakly supervised Semantic Segmentation Weakly-Supervised Semantic Segmentation
1 code implementation • 12 Mar 2024 • Chuangchuang Tan, Yao Zhao, Shikui Wei, Guanghua Gu, Ping Liu, Yunchao Wei
Consequently, these detectors have exhibited a lack of proficiency in learning the frequency domain and tend to overfit to the artifacts present in the training data, leading to suboptimal performance on unseen sources.
no code implementations • 12 Mar 2024 • Runmin Cong, Ronghui Sheng, Hao Wu, Yulan Guo, Yunchao Wei, WangMeng Zuo, Yao Zhao, Sam Kwong
On the one hand, the low-level detail embedding module is designed to supplement high-frequency color information of depth features in a residual mask manner at the low-level stages.
1 code implementation • 11 Mar 2024 • Chuangchuang Tan, Ping Liu, Renshuai Tao, Huan Liu, Yao Zhao, Baoyuan Wu, Yunchao Wei
Due to its unbias towards both the training and test sources, we define it as Data-Independent Operator (DIO) to achieve appealing improvements on unseen sources.
2 code implementations • 1 Mar 2024 • Yuxi Liu, Wenhan Yang, Huihui Bai, Yunchao Wei, Yao Zhao
However, there is no prior research on neural transform that focuses on specific regions.
Ranked #1 on Image Compression on kodak
no code implementations • 28 Dec 2023 • Yuyang Yin, Dejia Xu, Zhangyang Wang, Yao Zhao, Yunchao Wei
Our pipeline facilitates controllable 4D generation, enabling users to specify the motion via monocular video or adopt image-to-video generations, thus offering superior control over content creation.
1 code implementation • CVPR 2024 • Huan Liu, Zichang Tan, Chuangchuang Tan, Yunchao Wei, Yao Zhao, Jingdong Wang
In this paper, we study the problem of generalizable synthetic image detection, aiming to detect forgery images from diverse generative methods, e. g., GANs and diffusion models.
1 code implementation • NeurIPS 2023 • Mengyu Wang, Henghui Ding, Jun Hao Liew, Jiajun Liu, Yao Zhao, Yunchao Wei
We propose a model-agnostic solution called SegRefiner, which offers a novel perspective on this problem by interpreting segmentation refinement as a data generation process.
2 code implementations • CVPR 2024 • Chuangchuang Tan, Huan Liu, Yao Zhao, Shikui Wei, Guanghua Gu, Ping Liu, Yunchao Wei
Recently, the proliferation of highly realistic synthetic images, facilitated through a variety of GANs and Diffusions, has significantly heightened the susceptibility to misuse.
1 code implementation • 10 Dec 2023 • Yihan Hu, Yiheng Lin, Wei Wang, Yao Zhao, Yunchao Wei, Humphrey Shi
However, the presence of high computational overhead and the inconsistency of noise sampling between the training and inference processes pose significant obstacles to achieving this goal.
Ranked #1 on Image Matting on Distinctions-646
1 code implementation • CVPR 2024 • Zhongwei Ren, Zhicheng Huang, Yunchao Wei, Yao Zhao, Dongmei Fu, Jiashi Feng, Xiaojie Jin
PixelLM excels across various pixel-level image reasoning and understanding tasks, outperforming well-established methods in multiple benchmarks, including MUSE, single- and multi-referring segmentation.
1 code implementation • ICCV 2023 • Zekang Zhang, Guangyu Gao, Jianbo Jiao, Chi Harold Liu, Yunchao Wei
However, most state-of-the-art methods use the freeze strategy for stability, which compromises the model's plasticity. In contrast, releasing parameter training for plasticity could lead to the best performance for all categories, but this requires discriminative feature representation. Therefore, we prioritize the model's plasticity and propose the Contrast inter- and intra-class representations for Incremental Segmentation (CoinSeg), which pursues discriminative representations for flexible parameter tuning.
2 code implementations • NeurIPS 2023 • Siyu Jiao, Yunchao Wei, YaoWei Wang, Yao Zhao, Humphrey Shi
However, in the paper, we reveal that CLIP is insensitive to different mask proposals and tends to produce similar predictions for various mask proposals of the same image.
Open Vocabulary Semantic Segmentation Zero Shot Segmentation
no code implementations • 18 Sep 2023 • Huan Liu, Zichang Tan, Qiang Chen, Yunchao Wei, Yao Zhao, Jingdong Wang
Moreover, to address the semantic conflicts between image and frequency domains, the forgery-aware mutual module is developed to further enable the effective interaction of disparate image and frequency features, resulting in aligned and comprehensive visual forgery representations.
1 code implementation • 20 Aug 2023 • Yanda Li, Chi Zhang, Gang Yu, Zhibin Wang, Bin Fu, Guosheng Lin, Chunhua Shen, Ling Chen, Yunchao Wei
However, these datasets often exhibit domain bias, potentially constraining the generative capabilities of the models.
Ranked #116 on Visual Question Answering on MM-Vet
1 code implementation • 14 Aug 2023 • Hongguang Zhu, Yunchao Wei, Xiaodan Liang, Chunjie Zhang, Yao Zhao
Regarding the growing nature of real-world data, such an offline training paradigm on ever-expanding data is unsustainable, because models lack the continual learning ability to accumulate knowledge constantly.
no code implementations • 13 Aug 2023 • Yuyang Yin, Dejia Xu, Chuangchuang Tan, Ping Liu, Yao Zhao, Yunchao Wei
Low light enhancement has gained increasing importance with the rapid development of visual creation and editing.
1 code implementation • 3 Apr 2023 • Yanda Li, Zilong Huang, Gang Yu, Ling Chen, Yunchao Wei, Jianbo Jiao
The pre-training task is designed in a similar manner as image matting, where random trimap and alpha matte are generated to achieve an image disentanglement objective.
1 code implementation • CVPR 2023 • Man Liu, Feng Li, Chunjie Zhang, Yunchao Wei, Huihui Bai, Yao Zhao
Generalized Zero-Shot Learning (GZSL) identifies unseen categories by knowledge transferred from the seen domain, relying on the intrinsic interactions between visual and semantic information.
1 code implementation • ICCV 2023 • Kunyang Han, Yong liu, Jun Hao Liew, Henghui Ding, Yunchao Wei, Jiajun Liu, Yitong Wang, Yansong Tang, Yujiu Yang, Jiashi Feng, Yao Zhao
Recent advancements in pre-trained vision-language models, such as CLIP, have enabled the segmentation of arbitrary concepts solely from textual inputs, a process commonly referred to as open-vocabulary semantic segmentation (OVS).
Knowledge Distillation Open Vocabulary Semantic Segmentation +4
2 code implementations • ICCV 2023 • Gengwei Zhang, Liyuan Wang, Guoliang Kang, Ling Chen, Yunchao Wei
The goal of continual learning is to improve the performance of recognition models in learning sequentially arrived data.
1 code implementation • CVPR 2023 • Chuangchuang Tan, Yao Zhao, Shikui Wei, Guanghua Gu, Yunchao Wei
The key of fake image detection is to develop a generalized representation to describe the artifacts produced by generation models.
1 code implementation • CVPR 2023 • Mengxue Qu, Yu Wu, Yunchao Wei, Wu Liu, Xiaodan Liang, Yao Zhao
Extensive experiments show that our model achieves 52. 06% in terms of accuracy (versus 58. 93% in fully supervised setting) on RefCOCO+@testA, when only using 1% of the mask annotations.
no code implementations • CVPR 2023 • Guangrui Li, Guoliang Kang, Xiaohan Wang, Yunchao Wei, Yi Yang
With the help of adversarial training, the masking module can learn to generate source masks to mimic the pattern of irregular target noise, thereby narrowing the domain gap.
1 code implementation • ICCV 2023 • Hongguang Zhu, Yunchao Wei, Xiaodan Liang, Chunjie Zhang, Yao Zhao
Regarding the growing nature of real-world data, such an offline training paradigm on ever-expanding data is unsustainable, because models lack the continual learning ability to accumulate knowledge constantly.
no code implementations • ICCV 2023 • Yan Fang, Feng Zhu, Bowen Cheng, Luoqi Liu, Yao Zhao, Yunchao Wei
This work shows that locating the patch-wise noisy region is a better way to deal with noise.
1 code implementation • 5 Dec 2022 • Siyu Jiao, Gengwei Zhang, Shant Navasardyan, Ling Chen, Yao Zhao, Yunchao Wei, Humphrey Shi
Typical methods follow the paradigm to firstly learn prototypical features from support images and then match query features in pixel-level to obtain segmentation results.
1 code implementation • 13 Nov 2022 • Zekang Zhang, Guangyu Gao, Zhiyuan Fang, Jianbo Jiao, Yunchao Wei
Our MicroSeg is based on the assumption that background regions with strong objectness possibly belong to those concepts in the historical or future stages.
Class-Incremental Semantic Segmentation Continual Learning +1
1 code implementation • 26 Aug 2022 • Jiachen Li, Vidit Goel, Marianna Ohanyan, Shant Navasardyan, Yunchao Wei, Humphrey Shi
In this paper, we propose VMFormer: a transformer-based end-to-end method for video matting.
1 code implementation • 5 Aug 2022 • Feng Zhu, Zongxin Yang, Xin Yu, Yi Yang, Yunchao Wei
In this work, we propose a new online VIS paradigm named Instance As Identity (IAI), which models temporal information for both detection and tracking in an efficient way.
1 code implementation • 27 Jul 2022 • Mengxue Qu, Yu Wu, Wu Liu, Qiqi Gong, Xiaodan Liang, Olga Russakovsky, Yao Zhao, Yunchao Wei
Particularly, SiRi conveys a significant principle to the research of visual grounding, i. e., a better initialized vision-language encoder would help the model converge to a better local minimum, advancing the performance accordingly.
no code implementations • 17 Jun 2022 • Xiao Dong, Xunlin Zhan, Yunchao Wei, XiaoYong Wei, YaoWei Wang, Minlong Lu, Xiaochun Cao, Xiaodan Liang
Our goal in this research is to study a more realistic environment in which we can conduct weakly-supervised multi-modal instance-level product retrieval for fine-grained product categories.
1 code implementation • 12 Jun 2022 • Kang Liao, Chunyu Lin, Yunchao Wei, Yao Zhao
For the distortion synthesis, we propose a spiral distortion-aware perception module, in which the learning path keeps consistent with the distortion prior of the fisheye image.
1 code implementation • 18 Apr 2022 • Kang Liao, Xiangyu Xu, Chunyu Lin, Wenqi Ren, Yunchao Wei, Yao Zhao
Motivated by this analysis, we present a Cylin-Painting framework that involves meaningful collaborations between inpainting and outpainting and efficiently fuses the different arrangements, with a view to leveraging their complementary benefits on a seamless cylinder.
1 code implementation • CVPR 2022 • Peng-Tao Jiang, YuQi Yang, Qibin Hou, Yunchao Wei
Our framework conducts the global network to learn the captured rich object detail knowledge from a global view and thereby produces high-quality attention maps that can be directly used as pseudo annotations for semantic segmentation networks.
Ranked #17 on Weakly-Supervised Semantic Segmentation on PASCAL VOC 2012 test (using extra training data)
2 code implementations • 22 Mar 2022 • Zongxin Yang, Jiaxu Miao, Yunchao Wei, Wenguan Wang, Xiaohan Wang, Yi Yang
This paper delves into the challenges of achieving scalable and effective multi-object modeling for semi-supervised Video Object Segmentation (VOS).
1 code implementation • CVPR 2022 • Jiaxu Miao, Xiaohan Wang, Yu Wu, Wei Li, Xu Zhang, Yunchao Wei, Yi Yang
In contrast, our large-scale VIdeo Panoptic Segmentation in the Wild (VIPSeg) dataset provides 3, 536 videos and 84, 750 frames with pixel-level panoptic annotations, covering a wide range of real-world scenarios and categories.
no code implementations • 11 Nov 2021 • Yutong Gao, Liqian Liang, Congyan Lang, Songhe Feng, Yidong Li, Yunchao Wei
In this work, we focus on Interactive Human Parsing (IHP), which aims to segment a human image into multiple human body parts with guidance from users' interactions.
no code implementations • CVPR 2022 • Xiao Dong, Xunlin Zhan, Yangxin Wu, Yunchao Wei, Michael C. Kampffmeyer, XiaoYong Wei, Minlong Lu, YaoWei Wang, Xiaodan Liang
Despite the potential of multi-modal pre-training to learn highly discriminative feature representations from complementary data modalities, current progress is being slowed by the lack of large-scale modality-diverse datasets.
1 code implementation • 26 Aug 2021 • Wuyang Chen, Xinyu Gong, Junru Wu, Yunchao Wei, Humphrey Shi, Zhicheng Yan, Yi Yang, Zhangyang Wang
This work targets designing a principled and unified training-free framework for Neural Architecture Search (NAS), with high performance, low cost, and in-depth interpretation.
no code implementations • 19 Aug 2021 • Yaxiong Wang, Yunchao Wei, Xueming Qian, Li Zhu, Yi Yang
Superpixel segmentation has recently seen important progress benefiting from the advances in differentiable deep learning.
1 code implementation • ICCV 2021 • Xunlin Zhan, Yangxin Wu, Xiao Dong, Yunchao Wei, Minlong Lu, Yichi Zhang, Hang Xu, Xiaodan Liang
In this paper, we investigate a more realistic setting that aims to perform weakly-supervised multi-modal instance-level product retrieval among fine-grained product categories.
3 code implementations • IEEE 2021 • Peng-Tao Jiang, Chang-Bin Zhang, Qibin Hou, Ming-Ming Cheng, Yunchao Wei
To evaluate the quality of the class activation maps produced by LayerCAM, we apply them to weakly-supervised object localization and semantic segmentation.
1 code implementation • 20 Jun 2021 • Yaxiong Wang, Yunchao Wei, Xueming Qian, Li Zhu, Yi Yang
We aim to tackle the challenging yet practical scenery image outpainting task in this work.
no code implementations • 20 Jun 2021 • Ping Liu, Yuewei Lin, Yang He, Yunchao Wei, Liangli Zhen, Joey Tianyi Zhou, Rick Siow Mong Goh, Jingen Liu
In this paper, we propose to utilize Automated Machine Learning to adaptively search a neural architecture for deepfake detection.
no code implementations • CVPR 2021 • Jiaxu Miao, Yunchao Wei, Yu Wu, Chen Liang, Guangrui Li, Yi Yang
To the best of our knowledge, our VSPW is the first attempt to tackle the challenging video scene parsing task in the wild by considering diverse scenarios.
1 code implementation • 8 Jun 2021 • Bingfeng Zhang, Jimin Xiao, Jianbo Jiao, Yunchao Wei, Yao Zhao
More importantly, our approach can be readily applied to bounding box supervised instance segmentation task or other weakly supervised semantic segmentation tasks, with state-of-the-art or comparable performance among almot all weakly supervised tasks on PASCAL VOC or COCO dataset.
Box-supervised Instance Segmentation Graph Neural Network +4
1 code implementation • CVPR 2021 • Guangrui Li, Guoliang Kang, Yi Zhu, Yunchao Wei, Yi Yang
To better exploit the intrinsic structure of the target domain, we propose Domain Consensus Clustering (DCC), which exploits the domain consensus knowledge to discover discriminative clusters on both common samples and private ones.
Ranked #4 on Partial Domain Adaptation on Office-31
2 code implementations • NeurIPS 2021 • Gengwei Zhang, Guoliang Kang, Yi Yang, Yunchao Wei
Directly performing cross-attention may aggregate these features from support to query and bias the query features.
Ranked #55 on Few-Shot Semantic Segmentation on PASCAL-5i (1-Shot)
2 code implementations • NeurIPS 2021 • Zongxin Yang, Yunchao Wei, Yi Yang
The state-of-the-art methods learn to decode features with a single positive object and thus have to match and segment each target separately under multi-object scenarios, consuming multiple times computing resources.
Ranked #2 on Video Object Segmentation on DAVIS 2017 (test-dev) (using extra training data)
no code implementations • 2 Jun 2021 • Chen Liang, Yu Wu, Tianfei Zhou, Wenguan Wang, Zongxin Yang, Yunchao Wei, Yi Yang
Referring video object segmentation (RVOS) aims to segment video objects with the guidance of natural language reference.
1 code implementation • 15 May 2021 • Si Liu, Tianrui Hui, Shaofei Huang, Yunchao Wei, Bo Li, Guanbin Li
In this paper, we propose a Cross-Modal Progressive Comprehension (CMPC) scheme to effectively mimic human behaviors and implement it as a CMPC-I (Image) module and a CMPC-V (Video) module to improve referring image and video segmentation models.
Ranked #7 on Referring Expression Segmentation on J-HMDB
no code implementations • 18 Mar 2021 • Qianyu Feng, Yunchao Wei, MingMing Cheng, Yi Yang
Visual grounding is a long-lasting problem in vision-language understanding due to its diversity and complexity.
no code implementations • ICCV 2021 • Yaxiong Wang, Yunchao Wei, Xueming Qian, Li Zhu, Yi Yang
However, simply applying a series of convolution operations with limited receptive fields can only implicitly perceive the relations between the pixel and its surrounding grids.
no code implementations • ICCV 2021 • Kang Liao, Chunyu Lin, Yunchao Wei, Feng Li, Shangrong Yang, Yao Zhao
To our knowledge, we are the first to tackle the challenging rectification via outpainting, and our curve-aware strategy can reach a rectification construction with complete content and regular shape.
no code implementations • NeurIPS 2020 • Peike Li, Yunchao Wei, Yi Yang
Concretely, by exploring the pair-wise and list-wise structures, we impose the relations of generated visual features to be consistent with their counterparts in the semantic word embedding space.
2 code implementations • 25 Nov 2020 • Chang-Bin Zhang, Peng-Tao Jiang, Qibin Hou, Yunchao Wei, Qi Han, Zhen Li, Ming-Ming Cheng
Experiments demonstrate that based on the same classification models, the proposed approach can effectively improve the classification performance on CIFAR-100, ImageNet, and fine-grained datasets.
1 code implementation • NeurIPS 2020 • Guoliang Kang, Yunchao Wei, Yi Yang, Yueting Zhuang, Alexander G. Hauptmann
The conventional solution to this task is to minimize the discrepancy between source and target to enable effective knowledge transfer.
Ranked #25 on Synthetic-to-Real Translation on SYNTHIA-to-Cityscapes
no code implementations • 17 Oct 2020 • Yunchao Wei, Shuai Zheng, Ming-Ming Cheng, Hang Zhao, LiWei Wang, Errui Ding, Yi Yang, Antonio Torralba, Ting Liu, Guolei Sun, Wenguan Wang, Luc van Gool, Wonho Bae, Junhyug Noh, Jinhwan Seo, Gunhee Kim, Hao Zhao, Ming Lu, Anbang Yao, Yiwen Guo, Yurong Chen, Li Zhang, Chuangchuang Tan, Tao Ruan, Guanghua Gu, Shikui Wei, Yao Zhao, Mariia Dobko, Ostap Viniavskyi, Oles Dobosevych, Zhendong Wang, Zhenyuan Chen, Chen Gong, Huanqing Yan, Jun He
The purpose of the Learning from Imperfect Data (LID) workshop is to inspire and facilitate the research in developing novel approaches that would harness the imperfect data and improve the data-efficiency during training.
1 code implementation • 13 Oct 2020 • Zongxin Yang, Yunchao Wei, Yi Yang
This paper investigates the principles of embedding learning to tackle the challenging semi-supervised video object segmentation.
1 code implementation • CVPR 2020 • Shaofei Huang, Tianrui Hui, Si Liu, Guanbin Li, Yunchao Wei, Jizhong Han, Luoqi Liu, Bo Li
In addition to the CMPC module, we further leverage a simple yet effective TGFE module to integrate the reasoned multimodal features from different levels with the guidance of textual information.
no code implementations • 19 Aug 2020 • Yaxiong Wang, Yunchao Wei, Xueming Qian, Li Zhu, Yi Yang
Skin lesion segmentation is a crucial step in the computer-aided diagnosis of dermoscopic images.
1 code implementation • ECCV 2020 • Xiaolin Zhang, Yunchao Wei, Yi Yang
We learn a feature center for each category and realize the global feature consistency by forcing the object features to approach class-specific centers.
no code implementations • 17 Jun 2020 • Yaxiong Wang, Yunchao Wei, Xueming Qian, Li Zhu, Yi Yang
In this work, we take the image outpainting one step forward by allowing users to harvest personal custom outpainting results using sketches as the guidance.
1 code implementation • 9 Jun 2020 • Xiaolin Zhang, Yunchao Wei, Yi Yang, Fei Wu
To fulfill the direct evaluation, we annotate pixel-level object masks on the ILSVRC validation set.
1 code implementation • 22 May 2020 • Prateek Shroff, Tianlong Chen, Yunchao Wei, Zhangyang Wang
In this paper, we tried to focus on these marginal differences to extract more representative features.
no code implementations • 18 May 2020 • Ping Liu, Yunchao Wei, Zibo Meng, Weihong Deng, Joey Tianyi Zhou, Yi Yang
However, the performance of the current state-of-the-art facial expression recognition (FER) approaches is directly related to the labeled data for training.
1 code implementation • 21 Apr 2020 • Mang Tik Chiu, Xingqian Xu, Kai Wang, Jennifer Hobbs, Naira Hovakimyan, Thomas S. Huang, Honghui Shi, Yunchao Wei, Zilong Huang, Alexander Schwing, Robert Brunner, Ivan Dozier, Wyatt Dozier, Karen Ghandilyan, David Wilson, Hyunseong Park, Junhee Kim, Sungho Kim, Qinghui Liu, Michael C. Kampffmeyer, Robert Jenssen, Arnt B. Salberg, Alexandre Barbosa, Rodrigo Trevisan, Bingchen Zhao, Shaozuo Yu, Siwei Yang, Yin Wang, Hao Sheng, Xiao Chen, Jingyi Su, Ram Rajagopal, Andrew Ng, Van Thong Huynh, Soo-Hyung Kim, In-Seop Na, Ujjwal Baid, Shubham Innani, Prasad Dutande, Bhakti Baheti, Sanjay Talbar, Jianyu Tang
The first Agriculture-Vision Challenge aims to encourage research in developing novel and effective algorithms for agricultural pattern recognition from aerial images, especially for the semantic segmentation task associated with our challenge dataset.
no code implementations • IEEE 2020 • Shuang Qiu, Yao Zhao, Jianbo Jiao, Yunchao Wei, Shikui Wei
To this end, we propose to train the referring image segmentation model in a generative adversarial fashion, which well addresses the distribution similarity problem.
3 code implementations • 14 Apr 2020 • Zhedong Zheng, Tao Ruan, Yunchao Wei, Yi Yang, Tao Mei
This stage relaxes the full alignment between the training and testing domains, as it is agnostic to the target vehicle domain.
Ranked #1 on Vehicle Re-Identification on VehicleID
no code implementations • 2 Apr 2020 • Zhonghao Wang, Yunchao Wei, Rogerior Feris, JinJun Xiong, Wen-mei Hwu, Thomas S. Huang, Humphrey Shi
A key challenge of this task is how to alleviate the data distribution discrepancy between the source and target domains, i. e. reducing domain shift.
no code implementations • CVPR 2020 • Jiaxu Miao, Yunchao Wei, Yi Yang
Interactive video object segmentation (iVOS) aims at efficiently harvesting high-quality segmentation masks of the target object in a video with user interactions.
Ranked #5 on Interactive Video Object Segmentation on DAVIS 2017 (AUC-J metric)
no code implementations • 30 Mar 2020 • Jianbo Jiao, Linchao Bao, Yunchao Wei, Shengfeng He, Honghui Shi, Rynson Lau, Thomas S. Huang
This can be naturally generalized to span multiple scales with a Laplacian pyramid representation of the input data.
1 code implementation • CVPR 2020 • Zhonghao Wang, Mo Yu, Yunchao Wei, Rogerio Feris, JinJun Xiong, Wen-mei Hwu, Thomas S. Huang, Humphrey Shi
We consider the problem of unsupervised domain adaptation for semantic segmentation by easing the domain shift between the source domain (synthetic data) and the target domain (real data) in this work.
Ranked #8 on Semantic Segmentation on DensePASS
2 code implementations • ECCV 2020 • Zongxin Yang, Yunchao Wei, Yi Yang
This paper investigates the principles of embedding learning to tackle the challenging semi-supervised video object segmentation.
Ranked #8 on Video Object Segmentation on YouTube-VOS 2019
3 code implementations • 27 Feb 2020 • Zhedong Zheng, Yunchao Wei, Yi Yang
To our knowledge, University-1652 is the first drone-based geo-localization dataset and enables two new tasks, i. e., drone-view target localization and drone navigation.
Ranked #6 on Drone navigation on University-1652
1 code implementation • 24 Feb 2020 • Zilong Huang, Yunchao Wei, Xinggang Wang, Wenyu Liu, Thomas S. Huang, Humphrey Shi
Aggregating features in terms of different convolutional blocks or contextual embeddings has been proven to be an effective way to strengthen feature representations for semantic segmentation.
2 code implementations • CVPR 2020 • Mang Tik Chiu, Xingqian Xu, Yunchao Wei, Zilong Huang, Alexander Schwing, Robert Brunner, Hrant Khachatrian, Hovnatan Karapetyan, Ivan Dozier, Greg Rose, David Wilson, Adrian Tudor, Naira Hovakimyan, Thomas S. Huang, Honghui Shi
To encourage research in computer vision for agriculture, we present Agriculture-Vision: a large-scale aerial farmland image dataset for semantic segmentation of agricultural patterns.
1 code implementation • 19 Nov 2019 • Bingfeng Zhang, Jimin Xiao, Yunchao Wei, Ming-Jie Sun, Kai-Zhu Huang
Such reliable regions are then directly served as ground-truth labels for the parallel segmentation branch, where a newly designed dense energy loss function is adopted for optimization.
Ranked #22 on Semantic Segmentation on PASCAL VOC 2012 val
2 code implementations • 22 Oct 2019 • Peike Li, Yunqiu Xu, Yunchao Wei, Yi Yang
To tackle the problem of learning with label noises, this work introduces a purification strategy, called Self-Correction for Human Parsing (SCHP), to progressively promote the reliability of the supervised labels as well as the learned models.
Ranked #2 on Human Part Segmentation on PASCAL-Part
no code implementations • ICCV 2019 • Bowen Cheng, Liang-Chieh Chen, Yunchao Wei, Yukun Zhu, Zilong Huang, JinJun Xiong, Thomas Huang, Wen-mei Hwu, Honghui Shi
The multi-scale context module refers to the operations to aggregate feature responses from a large spatial extent, while the single-stage encoder-decoder structure encodes the high-level semantic information in the encoder path and recovers the boundary information in the decoder path.
4 code implementations • ICCV 2019 • Zilong Huang, Xinggang Wang, Yunchao Wei, Lichao Huang, Humphrey Shi, Wenyu Liu, Thomas S. Huang
Compared with the non-local block, the proposed recurrent criss-cross attention module requires 11x less GPU memory usage.
Ranked #7 on Semantic Segmentation on FoodSeg103 (using extra training data)
1 code implementation • ICCV 2019 • Yang Fu, Yunchao Wei, Guanshuo Wang, Yuqian Zhou, Honghui Shi, Thomas Huang
Upon our SSG, we further introduce a clustering-guided semisupervised approach named SSG ++ to conduct the one-shot domain adaption in an open set setting (i. e. the number of independent identities from the target domain is unknown).
no code implementations • 23 Nov 2018 • Bowen Cheng, Yunchao Wei, Jiahui Yu, Shiyu Chang, JinJun Xiong, Wen-mei Hwu, Thomas S. Huang, Humphrey Shi
While training on samples drawn from independent and identical distribution has been a de facto paradigm for optimizing image classification networks, humans learn new concepts in an easy-to-hard manner and on the selected examples progressively.
no code implementations • 9 Nov 2018 • Yang Fu, Xiaoyang Wang, Yunchao Wei, Thomas Huang
Thus, a more robust clip-level feature representation can be generated according to a weighted sum operation guided by the mined 2-D attention score matrix.
Large-Scale Person Re-Identification Video-Based Person Re-Identification
no code implementations • 6 Nov 2018 • Rui Qian, Yunchao Wei, Honghui Shi, Jiachen Li, Jiaying Liu, Thomas Huang
Semantic scene parsing is suffering from the fact that pixel-level annotations are hard to be collected.
no code implementations • NeurIPS 2018 • Qibin Hou, Peng-Tao Jiang, Yunchao Wei, Ming-Ming Cheng
To test the quality of the generated attention maps, we employ the mined object regions as heuristic cues for learning semantic segmentation models.
1 code implementation • 22 Oct 2018 • Xiaolin Zhang, Yunchao Wei, Yi Yang, Thomas Huang
In this way, the possibilities embedded in the produced similarity maps can be adapted to guide the process of segmenting objects.
Ranked #92 on Few-Shot Semantic Segmentation on PASCAL-5i (5-Shot)
3 code implementations • 5 Oct 2018 • Bowen Cheng, Yunchao Wei, Rogerio Feris, JinJun Xiong, Wen-mei Hwu, Thomas Huang, Humphrey Shi
In particular, DCR places a separate classification network in parallel with the localization network (base detector).
2 code implementations • 17 Sep 2018 • Tao Ruan, Ting Liu, Zilong Huang, Yunchao Wei, Shikui Wei, Yao Zhao, Thomas Huang
Human parsing has received considerable interest due to its wide application potentials.
Ranked #2 on Person Re-Identification on Market-1501-C
1 code implementation • ECCV 2018 • Xiaolin Zhang, Yunchao Wei, Guoliang Kang, Yi Yang, Thomas Huang
A stagewise approach is proposed to incorporate high confident object regions to learn the SPG masks.
Ranked #1 on Weakly-Supervised Object Localization on ILSVRC 2015
no code implementations • ECCV 2018 • Yunchao Wei, Zhiqiang Shen, Bowen Cheng, Honghui Shi, JinJun Xiong, Jiashi Feng, Thomas Huang
This work provides a simple approach to discover tight object bounding boxes with only image-level supervision, called Tight box mining with Surrounding Segmentation Context (TS2C).
no code implementations • CVPR 2018 • Yunchao Wei, Huaxin Xiao, Honghui Shi, Zequn Jie, Jiashi Feng, Thomas S. Huang
Despite remarkable progress, weakly supervised segmentation methods are still inferior to their fully supervised counterparts.
no code implementations • CVPR 2018 • Yunchao Wei, Huaxin Xiao, Honghui Shi, Zequn Jie, Jiashi Feng, Thomas S. Huang
It can produce dense and reliable object localization maps and effectively benefit both weakly- and semi- supervised semantic segmentation.
2 code implementations • CVPR 2018 • Xiaolin Zhang, Yunchao Wei, Jiashi Feng, Yi Yang, Thomas Huang
With such an adversarial learning, the two parallel-classifiers are forced to leverage complementary object regions for classification and can finally generate integral object localization together.
Ranked #2 on Weakly-Supervised Object Localization on ILSVRC 2016
1 code implementation • 14 Apr 2018 • Yang Fu, Yunchao Wei, Yuqian Zhou, Honghui Shi, Gao Huang, Xinchao Wang, Zhiqiang Yao, Thomas Huang
Despite the remarkable recent progress, person re-identification (Re-ID) approaches are still suffering from the failure cases where the discriminative body parts are missing.
Ranked #58 on Person Re-Identification on DukeMTMC-reID
no code implementations • CVPR 2018 • Zequn Jie, Pengfei Wang, Yonggen Ling, Bo Zhao, Yunchao Wei, Jiashi Feng, Wei Liu
Left-right consistency check is an effective way to enhance the disparity estimation by referring to the information from the opposite view.
7 code implementations • ECCV 2018 • Bowen Cheng, Yunchao Wei, Honghui Shi, Rogerio Feris, JinJun Xiong, Thomas Huang
Recent region-based object detectors are usually built with separate classification and localization branches on top of shared feature extraction networks.
no code implementations • 18 Nov 2017 • Huaxin Xiao, Yunchao Wei, Yu Liu, Maojun Zhang, Jiashi Feng
The performance of deep learning based semantic segmentation models heavily depends on sufficient data with careful annotations.
no code implementations • 4 Oct 2017 • Xiaodan Liang, Yunchao Wei, Liang Lin, Yunpeng Chen, Xiaohui Shen, Jianchao Yang, Shuicheng Yan
An intuition on human segmentation is that when a human is moving in a video, the video-context (e. g., appearance and motion clues) may potentially infer reasonable mask information for the whole human body.
no code implementations • ICCV 2017 • Jun Hao Liew, Yunchao Wei, Wei Xiong, Sim-Heng Ong, Jiashi Feng
The interactive image segmentation model allows users to iteratively add new inputs for refinement until a satisfactory result is finally obtained.
Ranked #10 on Interactive Segmentation on SBD (NoC@85 metric)
no code implementations • 18 Aug 2017 • Huaxin Xiao, Jiashi Feng, Yunchao Wei, Maojun Zhang
Through visualizing the differences, we can interpret the capability of different deep neural networks based saliency detection models and demonstrate that our proposed model indeed uses more reasonable structure for salient object detection.
no code implementations • CVPR 2017 • Jianan Li, Xiaodan Liang, Yunchao Wei, Tingfa Xu, Jiashi Feng, Shuicheng Yan
In this work, we address the small object detection problem by developing a single architecture that internally lifts representations of small objects to "super-resolved" ones, achieving similar characteristics as large objects and thus more discriminative for detection.
2 code implementations • 19 May 2017 • Jianshu Li, Jian Zhao, Yunchao Wei, Congyan Lang, Yidong Li, Terence Sim, Shuicheng Yan, Jiashi Feng
To address the multi-human parsing problem, we introduce a new multi-human parsing (MHP) dataset and a novel multi-human parsing model named MH-Parser.
Ranked #3 on Multi-Human Parsing on MHP v1.0
no code implementations • 16 May 2017 • Jimin Xiao, Yanchun Xie, Tammam Tillo, Kai-Zhu Huang, Yunchao Wei, Jiashi Feng
In addition, to relieve the negative effect caused by varying visual appearances of the same individual, IAN introduces a novel center loss that can increase the intra-class compactness of feature representations.
no code implementations • CVPR 2017 • Zequn Jie, Yunchao Wei, Xiaojie Jin, Jiashi Feng, Wei Liu
To overcome this issue, we propose a deep self-taught learning approach, which makes the detector learn the object-level features reliable for acquiring tight positive samples and afterwards re-train itself based on them.
no code implementations • CVPR 2017 • Yunchao Wei, Jiashi Feng, Xiaodan Liang, Ming-Ming Cheng, Yao Zhao, Shuicheng Yan
We investigate a principle way to progressively mine discriminative object regions using classification networks to address the weakly-supervised semantic segmentation problems.
no code implementations • 7 Dec 2016 • Qinbin Hou, Puneet Kumar Dokania, Daniela Massiceti, Yunchao Wei, Ming-Ming Cheng, Philip Torr
We focus on the following three aspects of EM: (i) initialization; (ii) latent posterior estimation (E-step) and (iii) the parameter update (M-step).
Weakly supervised Semantic Segmentation Weakly-Supervised Semantic Segmentation