no code implementations • ECCV 2020 • Lijun Wang, Jianming Zhang, Yifan Wang, Huchuan Lu, Xiang Ruan
This paper proposes a hierarchical loss for monocular depth estimation, which measures the differences between the prediction and ground truth in hierarchical embedding spaces of depth maps.
1 code implementation • 28 Apr 2024 • Haiwen Diao, Ying Zhang, Shang Gao, Xiang Ruan, Huchuan Lu
Specifically, we propose a brand-new Deep Boosting Learning (DBL) algorithm, where an anchor branch is first trained to provide insights into the data properties, with a target branch gaining more advanced knowledge to develop optimal features and distance metrics.
1 code implementation • 23 Mar 2023 • Haiwen Diao, Ying Zhang, Wei Liu, Xiang Ruan, Huchuan Lu
Exploiting fine-grained correspondence and visual-semantic alignments has shown great potential in image-text matching.
Ranked #2 on Image Retrieval on Flickr30K 1K test
1 code implementation • CVPR 2022 • Pengyu Zhang, Jie Zhao, Dong Wang, Huchuan Lu, Xiang Ruan
With the popularity of multi-modal sensors, visible-thermal (RGB-T) object tracking is to achieve robust performance and wider application scenarios with the guidance of objects' temperature information.
Ranked #8 on Rgb-T Tracking on GTOT
1 code implementation • 25 Mar 2022 • Xin Chen, Bin Yan, Jiawen Zhu, Huchuan Lu, Xiang Ruan, Dong Wang
First, we present a transformer tracking (named TransT) method based on the Siamese-like feature extraction backbone, the designed attention-based fusion mechanism, and the classification and regression head.
1 code implementation • 29 Jan 2021 • Xiaoqi Zhao, Youwei Pang, Lihe Zhang, Huchuan Lu, Xiang Ruan
Existing CNNs-Based RGB-D salient object detection (SOD) networks are all required to be pretrained on the ImageNet to learn the hierarchy features which helps provide a good initialization.
no code implementations • CVPR 2018 • Tiantian Wang, Lihe Zhang, Shuo Wang, Huchuan Lu, Gang Yang, Xiang Ruan, Ali Borji
Moreover, to effectively recover object boundaries, we propose a local Boundary Refinement Network (BRN) to adaptively learn the local contextual information for each spatial position.
Ranked #15 on RGB Salient Object Detection on DUTS-TE
1 code implementation • ICCV 2017 • Pingping Zhang, Dong Wang, Huchuan Lu, Hongyu Wang, Xiang Ruan
In addition, to achieve accurate boundary inference and semantic enhancement, edge-aware feature maps in low-level layers and the predicted results of low resolution features are recursively embedded into the learning framework.
Ranked #22 on RGB Salient Object Detection on DUTS-TE (max F-measure metric)
no code implementations • CVPR 2017 • Lijun Wang, Huchuan Lu, Yifan Wang, Mengyang Feng, Dong Wang, Bao-Cai Yin, Xiang Ruan
In the second stage, FIN is fine-tuned with its predicted saliency maps as ground truth.
no code implementations • CVPR 2016 • Ying Zhang, Baohua Li, Huchuan Lu, Atshushi Irie, Xiang Ruan
Person re-identification addresses the problem of matching people across disjoint camera views and extensive efforts have been made to seek either the robust feature representation or the discriminative matching metrics.
no code implementations • CVPR 2015 • Na Tong, Huchuan Lu, Xiang Ruan, Ming-Hsuan Yang
Furthermore, we show that the proposed bootstrap learning approach can be easily applied to other bottom-up saliency models for significant improvement.
no code implementations • CVPR 2015 • Lijun Wang, Huchuan Lu, Xiang Ruan, Ming-Hsuan Yang
In the global search stage, the local saliency map together with global contrast and geometric information are used as global features to describe a set of object candidate regions.
no code implementations • CVPR 2013 • Chuan Yang, Lihe Zhang, Huchuan Lu, Xiang Ruan, Ming-Hsuan Yang
The saliency of the image elements is defined based on their relevances to the given seeds or queries.