no code implementations • 3 Apr 2024 • Cheng Zhao, Su Sun, Ruoyu Wang, Yuliang Guo, Jun-Jun Wan, Zhou Huang, Xinyu Huang, Yingjie Victor Chen, Liu Ren
Most 3D Gaussian Splatting (3D-GS) based methods for urban scenes initialize 3D Gaussians directly with 3D LiDAR points, which not only underutilizes LiDAR data capabilities but also overlooks the potential advantages of fusing LiDAR with camera data.
1 code implementation • CVPR 2023 • Zhou Huang, Hang Dai, Tian-Zhu Xiang, Shuo Wang, Huai-Xin Chen, Jie Qin, Huan Xiong
Vision transformers have recently shown strong global context modeling capabilities in camouflaged object detection.
no code implementations • 12 Dec 2022 • Han Wang, Zhou Huang, Ganmin Yin, Yi Bao, Xiao Zhou, Yong Gao
The geographically weighted regression (GWR) is an essential tool for estimating the spatial variation of relationships between dependent and independent variables in geographical contexts.
2 code implementations • 8 Dec 2022 • Xiaoshui Huang, Zhou Huang, Sheng Li, Wentao Qu, Tong He, Yuenan Hou, Yifan Zuo, Wanli Ouyang
These token embeddings are concatenated with a task token and fed into the frozen CLIP transformer to learn point cloud representation.
no code implementations • 3 Nov 2022 • Yuan Hu, Zhibin Wang, Zhou Huang, Yu Liu
Given a set of polygon queries, the model learns the relations among them and encodes context information from the image to predict the final set of building polygons with fixed vertex numbers.
no code implementations • 5 May 2022 • Han Wang, Zhou Huang, Xiao Zhou, Ganmin Yin, Yi Bao, Yi Zhang
The attention fusion module incorporates route features with movement features to create a better spatial-temporal embedding.
1 code implementation • 7 Feb 2022 • Zhou Huang, Tian-Zhu Xiang, Huai-Xin Chen, Hang Dai
To this end, in this paper, we propose a novel weakly-supervised salient object detection framework to predict the saliency of remote sensing images from sparse scribble annotations.
no code implementations • 10 Jul 2020 • Zhou Huang, Huai-Xin Chen, Tao Zhou, Yun-Zhi Yang, Bi-Yuan Liu
Our MCI-Net includes two key components: 1) a cross-modal feature learning network, which is used to learn the high-level features for the RGB images and depth cues, effectively enabling the correlations between the two sources to be exploited; and 2) a multi-level interactive integration network, which integrates multi-level cross-modal features to boost the SOD performance.
no code implementations • 6 Apr 2020 • Zhou Huang, Huai-Xin Chen, Tao Zhou, Yun-Zhi Yang, Chang-Yin Wang, Bi-Yuan Liu
Furthermore, by using the proposed joint saliency measure, a variety of saliency maps are generated based on the discriminant dictionary.
1 code implementation • 2 Jan 2020 • Lei Dong, Zhou Huang, Jiang Zhang, Yu Liu
Understanding quantitative relationships between urban elements is crucial for a wide range of applications.
Physics and Society