no code implementations • 2 May 2023 • Fangjian Lin, Yizhe Ma, Sitong Wu, Long Yu, Shengwei Tian
Recently Transformer has shown good performance in several vision tasks due to its powerful modeling capabilities.
no code implementations • 2 May 2023 • Fangjian Lin, Yizhe Ma, Shengwei Tian
We validate the effectiveness of our method on different datasets and models and surpass previous state-of-the-art methods.
no code implementations • 1 May 2023 • Yizhe Ma, Fangjian Lin, Sitong Wu, Shengwei Tian, Long Yu
We expect that our PRSeg can promote the development of MLP-based decoder in semantic segmentation.
1 code implementation • 26 Apr 2023 • Fangjian Lin, Jianlong Yuan, Sitong Wu, Fan Wang, Zhibin Wang
Interestingly, the ranking of these spatial token mixers also changes under our UniNeXt, suggesting that an excellent spatial token mixer may be stifled due to a suboptimal general architecture, which further shows the importance of the study on the general architecture of vision backbone.
no code implementations • 26 Mar 2022 • Fangjian Lin, Tianyi Wu, Sitong Wu, Shengwei Tian, Guodong Guo
In this work, we focus on fusing multi-scale features from Transformer-based backbones for semantic segmentation, and propose a Feature Selective Transformer (FeSeFormer), which aggregates features from all scales (or levels) for each query feature.
no code implementations • 23 Mar 2022 • Fangjian Lin, Zhanhao Liang, Sitong Wu, Junjun He, Kai Chen, Shengwei Tian
In previous deep-learning-based methods, semantic segmentation has been regarded as a static or dynamic per-pixel classification task, \textit{i. e.,} classify each pixel representation to a specific category.
1 code implementation • 8 Jun 2021 • Sitong Wu, Tianyi Wu, Fangjian Lin, Shengwei Tian, Guodong Guo
Transformers have shown impressive performance in various natural language processing and computer vision tasks, due to the capability of modeling long-range dependencies.