1 code implementation • 15 Oct 2024 • Andong Lu, jiacong Zhao, Chenglong Li, Yun Xiao, Bin Luo
To handle this issue, we take original RGB and TIR networks as the teachers, and distill their content knowledge into two student networks respectively by the style-content orthogonal feature decoupling scheme.
Ranked #3 on Rgb-T Tracking on RGBT210
no code implementations • 16 Aug 2024 • Andong Lu, Wanyu Wang, Chenglong Li, Jin Tang, Bin Luo
Existing RGBT tracking methods often design various interaction models to perform cross-modal fusion of each layer, but can not execute the feature interactions among all layers, which plays a critical role in robust multimodal representation, due to large computational burden.
no code implementations • 5 Aug 2024 • Yun Xiao, jiacong Zhao, Andong Lu, Chenglong Li, Yin Lin, Bing Yin, Cong Liu
Existing Transformer-based RGBT trackers achieve remarkable performance benefits by leveraging self-attention to extract uni-modal features and cross-attention to enhance multi-modal feature interaction and template-search correlation computation.
Ranked #6 on Rgb-T Tracking on GTOT
1 code implementation • 4 May 2024 • Andong Lu, Wanyu Wang, Chenglong Li, Jin Tang, Bin Luo
In particular, we design a fusion structure space based on the hierarchical attention network, each attention-based fusion unit corresponding to a fusion operation and a combination of these attention units corresponding to a fusion structure.
Ranked #5 on Rgb-T Tracking on RGBT234
no code implementations • 3 Jan 2024 • Dengdi Sun, Yajie Pan, Andong Lu, Chenglong Li, Bin Luo
We introduce independent dynamic template tokens to interact with the search region, embedding temporal information to address appearance changes, while also retaining the involvement of the initial static template tokens in the joint feature extraction process to ensure the preservation of the original reliable target appearance information that prevent deviations from the target appearance caused by traditional temporal updates.
Ranked #13 on Rgb-T Tracking on RGBT210
no code implementations • 25 Dec 2023 • Andong Lu, Tianrui Zha, Chenglong Li, Jin Tang, XiaoFeng Wang, Bin Luo
To perform effective collaborative modeling between image relighting and person ReID tasks, we integrate the multilevel feature interactions in CENet.
1 code implementation • 25 Dec 2023 • Andong Lu, jiacong Zhao, Chenglong Li, Jin Tang, Bin Luo
To address this challenge, we propose a novel invertible prompt learning approach, which integrates the content-preserving prompts into a well-trained tracking model to adapt to various modality-missing scenarios, for robust RGBT tracking.
1 code implementation • 31 Aug 2023 • Andong Lu, Zhang Zhang, Yan Huang, Yifan Zhang, Chenglong Li, Jin Tang, Liang Wang
The illumination enhancement branch first estimates an enhanced image from the nighttime image using a nonlinear curve mapping method and then extracts the enhanced features.
no code implementations • 14 Nov 2020 • Andong Lu, Chenglong Li, Yuqing Yan, Jin Tang, Bin Luo
In specific, we use the modified VGG-M as the generality adapter to extract the modality-shared target representations. To extract the modality-specific features while reducing the computational complexity, we design a modality adapter, which adds a small block to the generality adapter in each layer and each modality in a parallel manner.
Ranked #13 on Rgb-T Tracking on GTOT
no code implementations • 14 Nov 2020 • Andong Lu, Cun Qian, Chenglong Li, Jin Tang, Liang Wang
To deal with the tracking failure caused by sudden camera motion, which often occurs in RGBT tracking, we design a resampling strategy based on optical flow algorithms.
Ranked #26 on Rgb-T Tracking on RGBT234
no code implementations • ECCV 2020 • Chenglong Li, Lei Liu, Andong Lu, Qing Ji, Jin Tang
RGB and thermal source data suffer from both shared and specific challenges, and how to explore and exploit them plays a critical role to represent the target appearance in RGBT tracking.
Ranked #12 on Rgb-T Tracking on GTOT
no code implementations • 17 Jul 2019 • Chenglong Li, Andong Lu, Aihua Zheng, Zhengzheng Tu, Jin Tang
In a specific, the generality adapter is to extract shared object representations, the modality adapter aims at encoding modality-specific information to deploy their complementary advantages, and the instance adapter is to model the appearance properties and temporal variations of a certain object.