no code implementations • 5 Sep 2024 • Chenglizhao Chen, Xinyu Liu, Mengke Song, Luming Li, Xu Yu, Shanchen Pang
In short, current methods struggle to integrate low-level visual and high-level action features, leading to poor anomaly detection in varied and complex scenes.
no code implementations • 22 Jan 2024 • Zhenyu Wu, Fengmao Lv, Chenglizhao Chen, Aimin Hao, Shuo Li
Colorectal polyp segmentation (CPS), an essential problem in medical image analysis, has garnered growing research attention.
no code implementations • CVPR 2024 • Chong Peng, Pengfei Zhang, Yongyong Chen, Zhao Kang, Chenglizhao Chen, Qiang Cheng
In this paper we propose a novel concept factorization method that seeks factor matrices using a cross-order positive semi-definite neighbor graph which provides comprehensive and complementary neighbor information of the data.
1 code implementation • CVPR 2024 • Wenfeng Song, Xingliang Jin, Shuai Li, Chenglizhao Chen, Aimin Hao, Xia Hou, Ning li, Hong Qin
Our MCM-LDM's cornerstone lies in its ability first to disentangle and then intricately weave together motion's tripartite components: motion trajectory motion content and motion style.
no code implementations • CVPR 2024 • Wenfeng Song, Xinyu Zhang, Shuai Li, Yang Gao, Aimin Hao, Xia Hou, Chenglizhao Chen, Ning li, Hong Qin
To date the quest to rapidly and effectively produce human-object interaction (HOI) animations directly from textual descriptions stands at the forefront of computer vision research.
1 code implementation • 6 Dec 2023 • Mengke Song, Linfeng Li, Dunquan Wu, Wenfeng Song, Chenglizhao Chen
To conquer, this paper proposes a new paradigm for saliency ranking, which aims to completely focus on ranking salient objects by their "importance order".
4 code implementations • 23 May 2023 • Guotao Wang, Chenglizhao Chen, Aimin Hao, Hong Qin, Deng-Ping Fan
The main reason is that there always exist "blind zooms" when using HMD to collect fixations since the users cannot keep spinning their heads to explore the entire panoptic scene all the time.
no code implementations • 13 Dec 2022 • Zhenyu Wu, Lin Wang, Wei Wang, Qing Xia, Chenglizhao Chen, Aimin Hao, Shuo Li
This paper attempts to answer this unexplored question by proving a hypothesis: there is a point-labeled dataset where saliency models trained on it can achieve equivalent performance when trained on the densely annotated dataset.
1 code implementation • 25 Oct 2022 • Zhenyu Wu, Shuai Li, Chenglizhao Chen, Hong Qin, Aimin Hao
First, instead of using the vanilla convolution with fixed kernel sizes for the encoder design, we propose the dynamic pyramid convolution (DPConv), which dynamically selects the best-suited kernel sizes w. r. t.
1 code implementation • 25 Oct 2022 • Zhenyu Wu, Lin Wang, Wei Wang, Tengfei Shi, Chenglizhao Chen, Aimin Hao, Shuo Li
In this paper, we propose a novel yet effective method for SOD, coined SODGAN, which can generate infinite high-quality image-mask pairs requiring only a few labeled data, and these synthesized pairs can replace the human-labeled DUTS-TR to train any off-the-shelf SOD model.
1 code implementation • 2 Jul 2022 • Yujia Sun, Shuo Wang, Chenglizhao Chen, Tian-Zhu Xiang
Camouflaged object detection (COD), segmenting objects that are elegantly blended into their surroundings, is a valuable yet challenging task.
1 code implementation • 20 Jun 2022 • Chenglizhao Chen, Mengke Song, Wenfeng Song, Li Guo, Muwei Jian
Video saliency detection (VSD) aims at fast locating the most attractive objects/things/patterns in a given video clip.
no code implementations • 20 Jun 2022 • Chenglizhao Chen, Hengsen Wang, Yuming Fang, Chong Peng
The existing state-of-the-art (SOTA) video salient object detection (VSOD) models have widely followed short-term methodology, which dynamically determines the balance between spatial and temporal saliency fusion by solely considering the current consecutive limited frames.
no code implementations • 22 Apr 2022 • Chong Peng, Yiqun Zhang, Yongyong Chen, Zhao Kang, Chenglizhao Chen, Qiang Cheng
Nonnegative matrix factorization (NMF) has been widely studied in recent years due to its effectiveness in representing nonnegative data with parts-based representations.
no code implementations • 8 Jan 2022 • Chong Peng, Yang Liu, Yongyong Chen, Xinxin Wu, Andrew Cheng, Zhao Kang, Chenglizhao Chen, Qiang Cheng
In this paper, we propose a novel nonconvex approach to robust principal component analysis for HSI denoising, which focuses on simultaneously developing more accurate approximations to both rank and column-wise sparsity for the low-rank and sparse components, respectively.
1 code implementation • 27 Dec 2021 • Guotao Wang, Chenglizhao Chen, Deng-Ping Fan, Aimin Hao, Hong Qin
Moreover, we distill knowledge from these regions to obtain complete new spatial-temporal-audio (STA) fixation prediction (FP) networks, enabling broad applications in cases where video tags are not available.
1 code implementation • CVPR 2021 • Guotao Wang, Chenglizhao Chen, Deng-Ping Fan, Aimin Hao, Hong Qin
Thanks to the rapid advances in the deep learning techniques and the wide availability of large-scale training sets, the performances of video saliency detection models have been improving steadily and significantly.
1 code implementation • CVPR 2021 • Qiang Zhai, Xin Li, Fan Yang, Chenglizhao Chen, Hong Cheng, Deng-Ping Fan
Automatically detecting/segmenting object(s) that blend in with their surroundings is difficult for current models.
no code implementations • 3 Nov 2020 • Chong Peng, Qian Zhang, Zhao Kang, Chenglizhao Chen, Qiang Cheng
It directly uses 2D data as inputs such that the learning of representations benefits from inherent structures and relationships of the data.
no code implementations • 10 Aug 2020 • Zhen-Yu Wu, Shuai Li, Chenglizhao Chen, Aimin Hao, Hong Qin
In sharp contrast to the state-of-the-art (SOTA) methods that focus on learning pixel-wise saliency in "single image" using perceptual clues mainly, our method has investigated the "object-level semantic ranks between multiple images", of which the methodology is more consistent with the real human attention mechanism.
1 code implementation • 7 Aug 2020 • Chenglizhao Chen, Hongmeng Zhao, Huan Yang, Chong Peng, Teng Yu
The screen content images (SCIs) usually comprise various content types with sharp edges, in which the artifacts or distortions can be well sensed by the vanilla structure similarity measurement in a full reference manner.
1 code implementation • 7 Aug 2020 • Zhen-Yu Wu, Shuai Li, Chenglizhao Chen, Aimin Hao, Hong Qin
Finally, all these complementary multi-model deep features will be selectively fused to make high-performance salient object detections.
1 code implementation • 7 Aug 2020 • Chenglizhao Chen, Jipeng Wei, Chong Peng, Hong Qin
The existing fusion based RGB-D salient object detection methods usually adopt the bi-stream structure to strike the fusion trade-off between RGB and depth (D).
Ranked #17 on
RGB-D Salient Object Detection
on NJU2K
1 code implementation • 7 Aug 2020 • Chenglizhao Chen, Jia Song, Chong Peng, Guodong Wang, Yuming Fang
Consequently, we can achieve a significant performance improvement by using this new training set to start a new round of network training.
no code implementations • 7 Aug 2020 • Zhen-Yu Wu, Shuai Li, Chenglizhao Chen, Aimin Hao, Hong Qin
Compared with the conventional hand-crafted approaches, the deep learning based methods have achieved tremendous performance improvements by training exquisitely crafted fancy networks over large-scale training sets.
1 code implementation • 7 Aug 2020 • Xuehao Wang, Shuai Li, Chenglizhao Chen, Aimin Hao, Hong Qin
Previous RGB-D salient object detection (SOD) methods have widely adopted deep learning tools to automatically strike a trade-off between RGB and D (depth), whose key rationale is to take full advantage of their complementary nature, aiming for a much-improved SOD performance than that of using either of them solely.
1 code implementation • 7 Aug 2020 • Xuehao Wang, Shuai Li, Chenglizhao Chen, Yuming Fang, Aimin Hao, Hong Qin
Existing RGB-D salient object detection methods treat depth information as an independent component to complement its RGB part, and widely follow the bi-stream parallel network architecture.
1 code implementation • 7 Aug 2020 • Chenglizhao Chen, Guotao Wang, Chong Peng, Dingwen Zhang, Yuming Fang, Hong Qin
In this way, even though the overall video saliency quality is heavily dependent on its spatial branch, however, the performance of the temporal branch still matter.
1 code implementation • 2 Aug 2020 • Yunxiao Li, Shuai Li, Chenglizhao Chen, Aimin Hao, Hong Qin
With the rapid development of deep learning techniques, image saliency deep models trained solely by spatial information have occasionally achieved detection performance for video data comparable to that of the models trained by both spatial and temporal information.
no code implementations • 19 May 2020 • Chong Peng, Zhilu Zhang, Zhao Kang, Chenglizhao Chen, Qiang Cheng
In particular, projection matrices are sought under the guidance of building new data representations, such that the spatial information is retained and projections are enhanced by the goal of clustering, which helps construct optimal projection directions.
no code implementations • 9 Jul 2019 • Chong Peng, Zhao Kang, Chenglizhao Chen, Qiang Cheng
Existing nonnegative matrix factorization methods focus on learning global structure of the data to construct basis and coefficient matrices, which ignores the local structure that commonly exists among data.
no code implementations • CVPR 2019 • Chong Peng, Chenglizhao Chen, Zhao Kang, Jianbo Li, Qiang Cheng
This drawback has limited the application of RPCA in solving real world problems.