Search Results for author: Sifan Zhou

Found 19 papers, 12 papers with code

MambaQuant: Quantizing the Mamba Family with Variance Aligned Rotation Methods

1 code implementation23 Jan 2025 Zukang Xu, Yuxuan Yue, Xing Hu, Zhihang Yuan, Zixu Jiang, Zhixuan Chen, Jiangyong Yu, Chen Xu, Sifan Zhou, Dawei Yang

To these ends, we propose MambaQuant, a post-training quantization (PTQ) framework consisting of: 1) Karhunen-Loeve Transformation (KLT) enhanced rotation, rendering the rotation matrix adaptable to diverse channel distributions.

Mamba Quantization

OstQuant: Refining Large Language Model Quantization with Orthogonal and Scaling Transformations for Better Distribution Fitting

1 code implementation23 Jan 2025 Xing Hu, Yuan Cheng, Dawei Yang, Zukang Xu, Zhihang Yuan, Jiangyong Yu, Chen Xu, Zhe Jiang, Sifan Zhou

We complement QSUR with mathematical derivations that examine the effects and limitations of various transformations, guiding our development of Orthogonal and Scaling Transformation-based Quantization (OSTQuant).

Language Modeling Language Modelling +2

GSRender: Deduplicated Occupancy Prediction via Weakly Supervised 3D Gaussian Splatting

no code implementations19 Dec 2024 Qianpu Sun, Changyong Shu, Sifan Zhou, Zichen Yu, Yan Chen, Dawei Yang, Yuan Chun

Consequently, we propose GSRender, which naturally employs 3D Gaussian Splatting for occupancy prediction, simplifying the sampling process.

3D Reconstruction NeRF

PTQ4RIS: Post-Training Quantization for Referring Image Segmentation

1 code implementation25 Sep 2024 Xiaoyan Jiang, Hang Yang, Kaiying Zhu, Xihe Qiu, Shibo Zhao, Sifan Zhou

Specifically, we first conduct an in-depth analysis of the root causes of performance degradation in RIS model quantization and propose dual-region quantization (DRQ) and reorder-based outlier-retained quantization (RORQ) to address the quantization difficulties in visual and text encoders.

Image Segmentation Quantization +2

Sub-SA: Strengthen In-context Learning via Submodular Selective Annotation

1 code implementation8 Jul 2024 Jian Qian, Miao Sun, Sifan Zhou, Ziyu Zhao, Ruizhi Hun, Patrick Chiang

In Sub-SA, we design a submodular function that facilitates effective subset selection for annotation and demonstrates the characteristics of monotonically and submodularity from the theoretical perspective.

Diversity In-Context Learning

P2P: Part-to-Part Motion Cues Guide a Strong Tracking Framework for LiDAR Point Clouds

1 code implementation7 Jul 2024 Jiahao Nie, Fei Xie, Sifan Zhou, Xueyi Zhou, Dong-Kyu Chae, Zhiwei He

Moreover, under the same point-based representation, P2P-point outperforms the previous motion tracker M$^2$Track by \textbf{3. 3\%} and \textbf{6. 7\%} on the KITTI and NuScenes, while running at a considerably high speed of \textbf{107 Fps} on a single RTX3090 GPU.

3D Single Object Tracking Object Tracking

PillarHist: A Quantization-aware Pillar Feature Encoder based on Height-aware Histogram

no code implementations29 May 2024 Sifan Zhou, Zhihang Yuan, Dawei Yang, Xubin Wen, Xing Hu, Yuguang Shi, Ziyu Zhao, Xiaobo Lu

To address above issue, we first unveil the importance of different input information during PFE and identify the height dimension as a key factor in enhancing 3D detection performance.

3D Object Detection Autonomous Driving +2

I-LLM: Efficient Integer-Only Inference for Fully-Quantized Low-Bit Large Language Models

no code implementations28 May 2024 Xing Hu, Yuan Cheng, Dawei Yang, Zhihang Yuan, Jiangyong Yu, Chen Xu, Sifan Zhou

Post-training quantization (PTQ) serves as a potent technique to accelerate the inference of large language models (LLMs).

Quantization

WKVQuant: Quantizing Weight and Key/Value Cache for Large Language Models Gains More

no code implementations19 Feb 2024 Yuxuan Yue, Zhihang Yuan, Haojie Duanmu, Sifan Zhou, Jianlong Wu, Liqiang Nie

Large Language Models (LLMs) face significant deployment challenges due to their substantial memory requirements and the computational demands of auto-regressive text generation process.

Quantization Text Generation

LiDAR-PTQ: Post-Training Quantization for Point Cloud 3D Object Detection

1 code implementation29 Jan 2024 Sifan Zhou, Liang Li, Xinyu Zhang, Bo Zhang, Shipeng Bai, Miao Sun, Ziyu Zhao, Xiaobo Lu, Xiangxiang Chu

To our knowledge, for the very first time in lidar-based 3D detection tasks, the PTQ INT8 model's accuracy is almost the same as the FP32 model while enjoying $3\times$ inference speedup.

3D Object Detection Autonomous Vehicles +3

RobustCalib: Robust Lidar-Camera Extrinsic Calibration with Consistency Learning

no code implementations2 Dec 2023 Shuang Xu, Sifan Zhou, Zhi Tian, Jizhou Ma, Qiong Nie, Xiangxiang Chu

Current traditional methods for LiDAR-camera extrinsics estimation depend on offline targets and human efforts, while learning-based approaches resort to iterative refinement for calibration results, posing constraints on their generalization and application in on-board systems.

Real-time 3D Single Object Tracking with Transformer

1 code implementation2 Sep 2022 Jiayao Shan, Sifan Zhou, Yubo Cui, Zheng Fang

PTT module in the voting stage could model the interactions among point patches, which learns context-dependent features.

3D Single Object Tracking Autonomous Driving +2

3D Object Tracking with Transformer

1 code implementation28 Oct 2021 Yubo Cui, Zheng Fang, Jiayao Shan, Zuoxu Gu, Sifan Zhou

By using cross-attention, the transformer decoder fuses features and includes more target cues into the current point cloud feature to compute the region attentions, which makes the similarity computing more efficient.

3D Object Tracking Decoder +2

3D-SiamRPN: An End-to-End Learning Method for Real-Time 3D Single Object Tracking Using Raw Point Cloud

no code implementations12 Aug 2021 Zheng Fang, Sifan Zhou, Yubo Cui, Sebastian Scherer

Then, to fuse the information of features in the two branches and obtain their similarity, we propose two cross correlation modules, named Pointcloud-wise and Point-wise respectively.

3D Single Object Tracking Object +2

Cannot find the paper you are looking for? You can Submit a new open access paper.