no code implementations • 18 Feb 2025 • Sifan Zhou, Shuo Wang, Zhihang Yuan, Mingjia Shi, Yuzhang Shang, Dawei Yang
Large Language Models (LLMs) fine-tuning technologies have achieved remarkable results.
1 code implementation • 23 Jan 2025 • Zukang Xu, Yuxuan Yue, Xing Hu, Zhihang Yuan, Zixu Jiang, Zhixuan Chen, Jiangyong Yu, Chen Xu, Sifan Zhou, Dawei Yang
To these ends, we propose MambaQuant, a post-training quantization (PTQ) framework consisting of: 1) Karhunen-Loeve Transformation (KLT) enhanced rotation, rendering the rotation matrix adaptable to diverse channel distributions.
1 code implementation • 23 Jan 2025 • Xing Hu, Yuan Cheng, Dawei Yang, Zukang Xu, Zhihang Yuan, Jiangyong Yu, Chen Xu, Zhe Jiang, Sifan Zhou
We complement QSUR with mathematical derivations that examine the effects and limitations of various transformations, guiding our development of Orthogonal and Scaling Transformation-based Quantization (OSTQuant).
no code implementations • 19 Dec 2024 • Qianpu Sun, Changyong Shu, Sifan Zhou, Zichen Yu, Yan Chen, Dawei Yang, Yuan Chun
Consequently, we propose GSRender, which naturally employs 3D Gaussian Splatting for occupancy prediction, simplifying the sampling process.
1 code implementation • 3 Dec 2024 • Zhaofeng Hu, Sifan Zhou, Shibo Zhao, Zhihang Yuan
3D single object tracking is essential in autonomous driving and robotics.
1 code implementation • 25 Sep 2024 • Xiaoyan Jiang, Hang Yang, Kaiying Zhu, Xihe Qiu, Shibo Zhao, Sifan Zhou
Specifically, we first conduct an in-depth analysis of the root causes of performance degradation in RIS model quantization and propose dual-region quantization (DRQ) and reorder-based outlier-retained quantization (RORQ) to address the quantization difficulties in visual and text encoders.
1 code implementation • 8 Jul 2024 • Jian Qian, Miao Sun, Sifan Zhou, Ziyu Zhao, Ruizhi Hun, Patrick Chiang
In Sub-SA, we design a submodular function that facilitates effective subset selection for annotation and demonstrates the characteristics of monotonically and submodularity from the theoretical perspective.
1 code implementation • 7 Jul 2024 • Jiahao Nie, Fei Xie, Sifan Zhou, Xueyi Zhou, Dong-Kyu Chae, Zhiwei He
Moreover, under the same point-based representation, P2P-point outperforms the previous motion tracker M$^2$Track by \textbf{3. 3\%} and \textbf{6. 7\%} on the KITTI and NuScenes, while running at a considerably high speed of \textbf{107 Fps} on a single RTX3090 GPU.
no code implementations • 29 May 2024 • Sifan Zhou, Zhihang Yuan, Dawei Yang, Xubin Wen, Xing Hu, Yuguang Shi, Ziyu Zhao, Xiaobo Lu
To address above issue, we first unveil the importance of different input information during PFE and identify the height dimension as a key factor in enhancing 3D detection performance.
no code implementations • 28 May 2024 • Xing Hu, Yuan Cheng, Dawei Yang, Zhihang Yuan, Jiangyong Yu, Chen Xu, Sifan Zhou
Post-training quantization (PTQ) serves as a potent technique to accelerate the inference of large language models (LLMs).
1 code implementation • 11 Apr 2024 • Weisheng Xu, Sifan Zhou, Zhihang Yuan
LiDAR-based 3D single object tracking (3D SOT) is a critical issue in robotics and autonomous driving.
no code implementations • 19 Feb 2024 • Yuxuan Yue, Zhihang Yuan, Haojie Duanmu, Sifan Zhou, Jianlong Wu, Liqiang Nie
Large Language Models (LLMs) face significant deployment challenges due to their substantial memory requirements and the computational demands of auto-regressive text generation process.
1 code implementation • 29 Jan 2024 • Sifan Zhou, Liang Li, Xinyu Zhang, Bo Zhang, Shipeng Bai, Miao Sun, Ziyu Zhao, Xiaobo Lu, Xiangxiang Chu
To our knowledge, for the very first time in lidar-based 3D detection tasks, the PTQ INT8 model's accuracy is almost the same as the FP32 model while enjoying $3\times$ inference speedup.
no code implementations • 2 Dec 2023 • Shuang Xu, Sifan Zhou, Zhi Tian, Jizhou Ma, Qiong Nie, Xiangxiang Chu
Current traditional methods for LiDAR-camera extrinsics estimation depend on offline targets and human efforts, while learning-based approaches resort to iterative refinement for calibration results, posing constraints on their generalization and application in on-board systems.
1 code implementation • 5 Feb 2023 • Sifan Zhou, Zhi Tian, Xiangxiang Chu, Xinyu Zhang, Bo Zhang, Xiaobo Lu, Chengjian Feng, Zequn Jie, Patrick Yin Chiang, Lin Ma
The deployment of 3D detectors strikes one of the major challenges in real-world self-driving scenarios.
1 code implementation • 2 Sep 2022 • Jiayao Shan, Sifan Zhou, Yubo Cui, Zheng Fang
PTT module in the voting stage could model the interactions among point patches, which learns context-dependent features.
1 code implementation • 28 Oct 2021 • Yubo Cui, Zheng Fang, Jiayao Shan, Zuoxu Gu, Sifan Zhou
By using cross-attention, the transformer decoder fuses features and includes more target cues into the current point cloud feature to compute the region attentions, which makes the similarity computing more efficient.
1 code implementation • 14 Aug 2021 • Jiayao Shan, Sifan Zhou, Zheng Fang, Yubo Cui
Our code is open-sourced for the robotics community at https://github. com/shanjiayao/PTT.
no code implementations • 12 Aug 2021 • Zheng Fang, Sifan Zhou, Yubo Cui, Sebastian Scherer
Then, to fuse the information of features in the two branches and obtain their similarity, we propose two cross correlation modules, named Pointcloud-wise and Point-wise respectively.