Search Results for author: Xiaoyun Yang

Found 18 papers, 13 papers with code

Regression of Instance Boundary by Aggregated CNN and GCN

no code implementations ECCV 2020 Yanda Meng, Wei Meng, Dongxu Gao, Yitian Zhao, Xiaoyun Yang, Xiaowei Huang, Yalin Zheng

In particular, thanks to the proposed aggregation GCN, our network benefits from direct feature learning of the instances’ boundary locations and the spatial information propagation across the image.

Image Segmentation regression +1

3D Dense Face Alignment with Fused Features by Aggregating CNNs and GCNs

no code implementations9 Mar 2022 Yanda Meng, Xu Chen, Dongxu Gao, Yitian Zhao, Xiaoyun Yang, Yihong Qiao, Xiaowei Huang, Yalin Zheng

In this paper, we propose a novel multi-level aggregation network to regress the coordinates of the vertices of a 3D face from a single 2D image in an end-to-end manner.

3D Face Alignment 3D Face Reconstruction +1

Counting with Adaptive Auxiliary Learning

1 code implementation8 Mar 2022 Yanda Meng, Joshua Bridge, Meng Wei, Yitian Zhao, Yihong Qiao, Xiaoyun Yang, Xiaowei Huang, Yalin Zheng

This paper proposes an adaptive auxiliary task learning based approach for object counting problems.

Auxiliary Learning Object Counting

BI-GCN: Boundary-Aware Input-Dependent Graph Convolution Network for Biomedical Image Segmentation

1 code implementation27 Oct 2021 Yanda Meng, Hongrun Zhang, Dongxu Gao, Yitian Zhao, Xiaoyun Yang, Xuesheng Qian, Xiaowei Huang, Yalin Zheng

Our model is well-suited to obtain global semantic region information while also accommodates local spatial boundary characteristics simultaneously.

Image Segmentation Segmentation +1

Video Annotation for Visual Tracking via Selection and Refinement

1 code implementation ICCV 2021 Kenan Dai, Jie Zhao, Lijun Wang, Dong Wang, Jianhua Li, Huchuan Lu, Xuesheng Qian, Xiaoyun Yang

Deep learning based visual trackers entail offline pre-training on large volumes of video datasets with accurate bounding box annotations that are labor-expensive to achieve.

Visual Tracking

Spatial Uncertainty-Aware Semi-Supervised Crowd Counting

1 code implementation ICCV 2021 Yanda Meng, Hongrun Zhang, Yitian Zhao, Xiaoyun Yang, Xuesheng Qian, Xiaowei Huang, Yalin Zheng

Semi-supervised approaches for crowd counting attract attention, as the fully supervised paradigm is expensive and laborious due to its request for a large number of images of dense crowd scenarios and their annotations.

Crowd Counting

A Video Is Worth Three Views: Trigeminal Transformers for Video-based Person Re-identification

no code implementations5 Apr 2021 Xuehu Liu, Pingping Zhang, Chenyang Yu, Huchuan Lu, Xuesheng Qian, Xiaoyun Yang

To capture richer perceptions and extract more comprehensive video representations, in this paper we propose a novel framework named Trigeminal Transformers (TMT) for video-based person Re-ID.

Video-Based Person Re-Identification

Transformer Tracking

1 code implementation CVPR 2021 Xin Chen, Bin Yan, Jiawen Zhu, Dong Wang, Xiaoyun Yang, Huchuan Lu

The correlation operation is a simple fusion manner to consider the similarity between the template and the search region.

Visual Object Tracking Visual Tracking

Watching You: Global-guided Reciprocal Learning for Video-based Person Re-identification

1 code implementation CVPR 2021 Xuehu Liu, Pingping Zhang, Chenyang Yu, Huchuan Lu, Xiaoyun Yang

Specifically, we first propose a Global-guided Correlation Estimation (GCE) to generate feature correlation maps of local features and global features, which help to localize the high- and low-correlation regions for identifying the same person.

Feature Correlation Video-Based Person Re-Identification

Alpha-Refine: Boosting Tracking Performance by Precise Bounding Box Estimation

1 code implementation4 Jul 2020 Bin Yan, Dong Wang, Huchuan Lu, Xiaoyun Yang

In recent years, the multiple-stage strategy has become a popular trend for visual tracking.

Visual Tracking

Jointly Modeling Motion and Appearance Cues for Robust RGB-T Tracking

no code implementations4 Jul 2020 Pengyu Zhang, Jie Zhao, Dong Wang, Huchuan Lu, Xiaoyun Yang

In this study, we propose a novel RGB-T tracking framework by jointly modeling both appearance and motion cues.

Rgb-T Tracking

High-Performance Long-Term Tracking with Meta-Updater

2 code implementations CVPR 2020 Kenan Dai, Yunhua Zhang, Dong Wang, Jianhua Li, Huchuan Lu, Xiaoyun Yang

Most top-ranked long-term trackers adopt the offline-trained Siamese architectures, thus, they cannot benefit from great progress of short-term trackers with online update.

Visual Object Tracking Visual Tracking +1

Cooling-Shrinking Attack: Blinding the Tracker with Imperceptible Noises

1 code implementation CVPR 2020 Bin Yan, Dong Wang, Huchuan Lu, Xiaoyun Yang

An effective and efficient perturbation generator is trained with a carefully designed adversarial loss, which can simultaneously cool hot regions where the target exists on the heatmaps and force the predicted bounding box to shrink, making the tracked target invisible to trackers.

Adversarial Attack

GradNet: Gradient-Guided Network for Visual Object Tracking

2 code implementations ICCV 2019 Peixia Li, Bo-Yu Chen, Wanli Ouyang, Dong Wang, Xiaoyun Yang, Huchuan Lu

In this work, we propose a novel gradient-guided network to exploit the discriminative information in gradients and update the template in the siamese network through feed-forward and backward operations.

Ranked #3 on Visual Object Tracking on OTB-2015 (Precision metric)

Object Template Matching +2

'Skimming-Perusal' Tracking: A Framework for Real-Time and Robust Long-term Tracking

1 code implementation ICCV 2019 Bin Yan, Haojie Zhao, Dong Wang, Huchuan Lu, Xiaoyun Yang

In this work, we present a novel robust and real-time long-term tracking framework based on the proposed skimming and perusal modules.

Cascaded Context Pyramid for Full-Resolution 3D Semantic Scene Completion

no code implementations ICCV 2019 Pingping Zhang, Wei Liu, Yinjie Lei, Huchuan Lu, Xiaoyun Yang

To address these issues, in this work we propose a novel deep learning framework, named Cascaded Context Pyramid Network (CCPNet), to jointly infer the occupancy and semantic labels of a volumetric 3D scene from a single depth image.

Ranked #5 on 3D Semantic Scene Completion on NYUv2 (using extra training data)

3D Semantic Scene Completion

Cannot find the paper you are looking for? You can Submit a new open access paper.