Search Results for author: Bonan Li

Found 16 papers, 1 papers with code

Top-Down Compression: Revisit Efficient Vision Token Projection for Visual Instruction Tuning

no code implementations17 May 2025 Bonan Li, ZiCheng Zhang, Songhua Liu, Weihao Yu, Xinchao Wang

Visual instruction tuning aims to enable large language models to comprehend the visual world, with a pivotal challenge lying in establishing an effective vision-to-language projection.

CoSER: Towards Consistent Dense Multiview Text-to-Image Generator for 3D Creation

no code implementations CVPR 2025 Bonan Li, ZiCheng Zhang, Xingyi Yang, Xinchao Wang

To further enhance cross-view consistency and alleviate content drift, CoSER rapidly scan all views in spiral bidirectional manner to aware holistic information and then scores each point based on semantic material.

3D Generation Text to 3D

LokiTalk: Learning Fine-Grained and Generalizable Correspondences to Enhance NeRF-based Talking Head Synthesis

no code implementations29 Nov 2024 Tianqi Li, Ruobing Zheng, Bonan Li, ZiCheng Zhang, Meng Wang, Jingdong Chen, Ming Yang

Despite significant progress in talking head synthesis since the introduction of Neural Radiance Fields (NeRF), visual artifacts and high training costs persist as major obstacles to large-scale commercial adoption.

NeRF Transfer Learning

DR-BFR: Degradation Representation with Diffusion Models for Blind Face Restoration

no code implementations15 Nov 2024 Xinmin Qiu, Bonan Li, ZiCheng Zhang, Congying Han, Tiande Guo

DR-BFR comprises two modules: 1) Degradation Representation Module (DRM): This module extracts degradation representation with content-irrelevant features from LQ faces and estimates a reasonable distribution in the degradation space through contrastive learning and a specially designed LQ reconstruction.

Blind Face Restoration Contrastive Learning +1

Focus on Neighbors and Know the Whole: Towards Consistent Dense Multiview Text-to-Image Generator for 3D Creation

no code implementations23 Aug 2024 Bonan Li, ZiCheng Zhang, Xingyi Yang, Xinchao Wang

To further enhance cross-view consistency and alleviate content drift, CoSER rapidly scan all views in spiral bidirectional manner to aware holistic information and then scores each point based on semantic material.

3D Generation Text to 3D

Blaze3DM: Marry Triplane Representation with Diffusion for 3D Medical Inverse Problem Solving

no code implementations24 May 2024 Jia He, Bonan Li, Ge Yang, Ziwen Liu

Solving 3D medical inverse problems such as image restoration and reconstruction is crucial in modern medical field.

compressed sensing Computational Efficiency +3

BlazeBVD: Make Scale-Time Equalization Great Again for Blind Video Deflickering

no code implementations10 Mar 2024 Xinmin Qiu, Congying Han, ZiCheng Zhang, Bonan Li, Tiande Guo, Pingyu Wang, Xuecheng Nie

Developing blind video deflickering (BVD) algorithms to enhance video temporal consistency, is gaining importance amid the flourish of image processing and video generation.

Video Generation Video Temporal Consistency

Learning Dynamic Tetrahedra for High-Quality Talking Head Synthesis

1 code implementation CVPR 2024 ZiCheng Zhang, Ruobing Zheng, Ziwen Liu, Congying Han, Tianqi Li, Meng Wang, Tiande Guo, Jingdong Chen, Bonan Li, Ming Yang

Recent works in implicit representations, such as Neural Radiance Fields (NeRF), have advanced the generation of realistic and animatable head avatars from video sequences.

NeRF

General Method for Solving Four Types of SAT Problems

no code implementations27 Dec 2023 Anqi Li, Congying Han, Tiande Guo, Haoran Li, Bonan Li

We experimentally verify that this paradigm can prune the unnecessary search space to find the optimal Boolean assignments for the problem.

reinforcement-learning Reinforcement Learning +1

DiffBFR: Bootstrapping Diffusion Model Towards Blind Face Restoration

no code implementations8 May 2023 Xinmin Qiu, Congying Han, ZiCheng Zhang, Bonan Li, Tiande Guo, Xuecheng Nie

This design is implemented with two key components: 1) Identity Restoration Module (IRM) for preserving the face details in results.

Blind Face Restoration Denoising +1

StyO: Stylize Your Face in Only One-shot

no code implementations6 Mar 2023 Bonan Li, ZiCheng Zhang, Xuecheng Nie, Congying Han, Yinhan Hu, Xinmin Qiu, Tiande Guo

And it introduces a novel triple reconstruction loss to fine-tune the pre-trained LDM for encoding style and content into corresponding identifiers; 2) Fine-grained Content Controller (FCC) for the recombination phase.

Disentanglement One-Shot Face Stylization

DropKey for Vision Transformer

no code implementations CVPR 2023 Bonan Li, Yinhan Hu, Xuecheng Nie, Congying Han, Xiangjian Jiang, Tiande Guo, Luoqi Liu

Given exploration on the above three questions, we present the novel DropKey method that regards Key as the drop unit and exploits decreasing schedule for drop ratio, improving ViTs in a general way.

Human-Object Interaction Detection image-classification +3

Masked Reconstruction Contrastive Learning with Information Bottleneck Principle

no code implementations15 Nov 2022 Ziwen Liu, Bonan Li, Congying Han, Tiande Guo, Xuecheng Nie

In order to alleviate the discriminative information overfitting problem effectively, we employ the reconstruction task to regularize the discriminative task.

Contrastive Learning image-classification +3

DropKey

no code implementations4 Aug 2022 Bonan Li, Yinhan Hu, Xuecheng Nie, Congying Han, Xiangjian Jiang, Tiande Guo, Luoqi Liu

Given exploration on the above three questions, we present the novel DropKey method that regards Key as the drop unit and exploits decreasing schedule for drop ratio, improving ViTs in a general way.

Human-Object Interaction Detection image-classification +3

DFS: A Diverse Feature Synthesis Model for Generalized Zero-Shot Learning

no code implementations19 Mar 2021 Bonan Li, Xuecheng Nie, Congying Han

In this paper, we propose to enhance the generalizability of GZSL models via improving feature diversity of unseen classes.

Diversity Generalized Zero-Shot Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.