Search Results for author: Bingxin Xu

Found 6 papers, 2 papers with code

E-CAR: Efficient Continuous Autoregressive Image Generation via Multistage Modeling

no code implementations18 Dec 2024 Zhihang Yuan, Yuzhang Shang, Hanling Zhang, Tongcheng Fang, Rui Xie, Bingxin Xu, Yan Yan, Shengen Yan, Guohao Dai, Yu Wang

Our approach not only enhances computational efficiency but also aligns naturally with image generation principles by operating in continuous token space and following a hierarchical generation process from coarse to fine details.

Computational Efficiency Denoising +1

freePruner: A Training-free Approach for Large Multimodal Model Acceleration

no code implementations23 Nov 2024 Bingxin Xu, Yuzhang Shang, Yunhao Ge, Qian Lou, Yan Yan

Large Multimodal Models (LMMs) have demonstrated impressive capabilities in visual-language tasks but face significant deployment challenges due to their high computational demands.

Quantization Question Answering +2

Interpolating Video-LLMs: Toward Longer-sequence LMMs in a Training-free Manner

no code implementations19 Sep 2024 Yuzhang Shang, Bingxin Xu, Weitai Kang, Mu Cai, Yuheng Li, Zehao Wen, Zhen Dong, Kurt Keutzer, Yong Jae Lee, Yan Yan

In this paper, we first identify the primary challenges in interpolating Video-LLMs: (1) the video encoder and modality alignment projector are fixed, preventing the integration of additional frames into Video-LLMs, and (2) the LLM backbone is limited in its content length capabilities, which complicates the processing of an increased number of video tokens.

LLaVA-PruMerge: Adaptive Token Reduction for Efficient Large Multimodal Models

1 code implementation22 Mar 2024 Yuzhang Shang, Mu Cai, Bingxin Xu, Yong Jae Lee, Yan Yan

In response, we propose PruMerge, a novel adaptive visual token reduction strategy that significantly reduces the number of visual tokens without compromising the performance of LMMs.

Language Modelling Large Language Model +4

Causal-DFQ: Causality Guided Data-free Network Quantization

1 code implementation ICCV 2023 Yuzhang Shang, Bingxin Xu, Gaowen Liu, Ramana Kompella, Yan Yan

Inspired by the causal understanding, we propose the Causality-guided Data-free Network Quantization method, Causal-DFQ, to eliminate the reliance on data via approaching an equilibrium of causality-driven intervened distributions.

Data Free Quantization Neural Network Compression

Cannot find the paper you are looking for? You can Submit a new open access paper.