Search Results for author: Honggang Chen

Found 12 papers, 5 papers with code

Compression with Global Guidance: Towards Training-free High-Resolution MLLMs Acceleration

1 code implementation9 Jan 2025 Xuyang Liu, ZiMing Wang, Yuhang Han, Yingyao Wang, Jiale Yuan, Jun Song, Bo Zheng, Linfeng Zhang, Siteng Huang, Honggang Chen

Multimodal large language models (MLLMs) have attracted considerable attention due to their exceptional performance in visual content understanding and reasoning.

Rethinking Token Reduction in MLLMs: Towards a Unified Paradigm for Training-Free Acceleration

no code implementations26 Nov 2024 Yuhang Han, Xuyang Liu, Pengxiang Ding, Donglin Wang, Honggang Chen, Qingsen Yan, Siteng Huang

To accelerate the inference of heavy Multimodal Large Language Models (MLLMs), this study rethinks the current landscape of training-free token reduction research.

Token Reduction

M$^2$IST: Multi-Modal Interactive Side-Tuning for Efficient Referring Expression Comprehension

no code implementations1 Jul 2024 Xuyang Liu, Ting Liu, Siteng Huang, Yi Xin, Yue Hu, Quanjun Yin, Donglin Wang, Honggang Chen

With M$^2$IST, standard transformer-based REC methods present competitive or even superior performance compared to full fine-tuning, while utilizing only 2. 11\% of the tunable parameters, 39. 61\% of the GPU memory, and 63. 46\% of the fine-tuning time required for full fine-tuning.

Referring Expression Referring Expression Comprehension +1

Perception- and Fidelity-aware Reduced-Reference Super-Resolution Image Quality Assessment

no code implementations15 May 2024 Xinying Lin, Xuyang Liu, Hong Yang, Xiaohai He, Honggang Chen

In this letter, we attempt to evaluate the perceptual quality and reconstruction fidelity of SR images considering LR images and scale factors.

Image Quality Assessment Image Super-Resolution

DARA: Domain- and Relation-aware Adapters Make Parameter-efficient Tuning for Visual Grounding

1 code implementation10 May 2024 Ting Liu, Xuyang Liu, Siteng Huang, Honggang Chen, Quanjun Yin, Long Qin, Donglin Wang, Yue Hu

Specifically, we propose \textbf{DARA}, a novel PETL method comprising \underline{\textbf{D}}omain-aware \underline{\textbf{A}}dapters (DA Adapters) and \underline{\textbf{R}}elation-aware \underline{\textbf{A}}dapters (RA Adapters) for VG.

Relation Spatial Reasoning +2

VGDiffZero: Text-to-image Diffusion Models Can Be Zero-shot Visual Grounders

1 code implementation3 Sep 2023 Xuyang Liu, Siteng Huang, Yachen Kang, Honggang Chen, Donglin Wang

Large-scale text-to-image diffusion models have shown impressive capabilities for generative tasks by leveraging strong vision-language alignment from pre-training.

Visual Grounding

Real-World Single Image Super-Resolution: A Brief Review

1 code implementation3 Mar 2021 Honggang Chen, Xiaohai He, Linbo Qing, Yuanyuan Wu, Chao Ren, Ce Zhu

More specifically, this review covers the critical publically available datasets and assessment metrics for RSISR, and four major categories of RSISR methods, namely the degradation modeling-based RSISR, image pairs-based RSISR, domain translation-based RSISR, and self-learning-based RSISR.

Computational Efficiency Image Super-Resolution +2

DPW-SDNet: Dual Pixel-Wavelet Domain Deep CNNs for Soft Decoding of JPEG-Compressed Images

no code implementations27 May 2018 Honggang Chen, Xiaohai He, Linbo Qing, Shuhua Xiong, Truong Q. Nguyen

The pixel domain deep network takes the four downsampled versions of the compressed image to form a 4-channel input and outputs a pixel domain prediction, while the wavelet domain deep network uses the 1-level discrete wavelet transformation (DWT) coefficients to form a 4-channel input to produce a DWT domain prediction.

Blocking JPEG Artifact Correction

CISRDCNN: Super-resolution of compressed images using deep convolutional neural networks

no code implementations19 Sep 2017 Honggang Chen, Xiaohai He, Chao Ren, Linbo Qing, Qizhi Teng

Experiments on compressed images produced by JPEG (we take the JPEG as an example in this paper) demonstrate that the proposed CISRDCNN yields state-of-the-art SR performance on commonly used test images and imagesets.

Image Super-Resolution

Cannot find the paper you are looking for? You can Submit a new open access paper.