Search Results for author: Shaozhe Hao

Found 13 papers, 11 papers with code

Señorita-2M: A High-Quality Instruction-based Dataset for General Video Editing by Video Specialists

no code implementations10 Feb 2025 Bojia Zi, Penghui Ruan, Marco Chen, Xianbiao Qi, Shaozhe Hao, Shihao Zhao, Youze Huang, Bin Liang, Rong Xiao, Kam-Fai Wong

On the other hand, end-to-end methods, which rely on edited video pairs for training, offer faster inference speeds but often produce poor editing results due to a lack of high-quality training video pairs.

Video Editing Video Generation

Elucidating the design space of language models for image generation

1 code implementation21 Oct 2024 Xuantong Liu, Shaozhe Hao, Xianbiao Qi, Tianyang Hu, Jun Wang, Rong Xiao, Yuan YAO

However, considering the essential differences between text and image modalities, the design space of language models for image generation remains underexplored.

Image Generation Text Generation

BiGR: Harnessing Binary Latent Codes for Image Generation and Improved Visual Representation Capabilities

1 code implementation18 Oct 2024 Shaozhe Hao, Xuantong Liu, Xianbiao Qi, Shihao Zhao, Bojia Zi, Rong Xiao, Kai Han, Kwan-Yee K. Wong

We introduce BiGR, a novel conditional image generation model using compact binary latent codes for generative training, focusing on enhancing both generation and representation capabilities.

Conditional Image Generation Image Inpainting +3

CusConcept: Customized Visual Concept Decomposition with Diffusion Models

1 code implementation1 Oct 2024 Zhi Xu, Shaozhe Hao, Kai Han

In this paper, we study a new and challenging task, customized concept decomposition, wherein the objective is to leverage diffusion models to decompose a single image and generate visual concepts from various perspectives.

Text to Image Generation Text-to-Image Generation

ArtiFade: Learning to Generate High-quality Subject from Blemished Images

no code implementations CVPR 2025 Shuya Yang, Shaozhe Hao, Yukang Cao, Kwan-Yee K. Wong

In this paper, we introduce ArtiFade to tackle this issue and successfully generate high-quality artifact-free images from blemished datasets.

Text to Image Generation Text-to-Image Generation

ConceptExpress: Harnessing Diffusion Models for Single-image Unsupervised Concept Extraction

1 code implementation9 Jul 2024 Shaozhe Hao, Kai Han, Zhengyao Lv, Shihao Zhao, Kwan-Yee K. Wong

To achieve this, we present ConceptExpress that tackles UCE by unleashing the inherent capabilities of pretrained diffusion models in two aspects.

Text to Image Generation Text-to-Image Generation

Bridging Different Language Models and Generative Vision Models for Text-to-Image Generation

2 code implementations12 Mar 2024 Shihao Zhao, Shaozhe Hao, Bojia Zi, Huaizhe xu, Kwan-Yee K. Wong

In this paper, we explore this objective and propose LaVi-Bridge, a pipeline that enables the integration of diverse pre-trained language models and generative vision models for text-to-image generation.

Language Modelling Text to Image Generation +1

ViCo: Plug-and-play Visual Condition for Personalized Text-to-image Generation

1 code implementation1 Jun 2023 Shaozhe Hao, Kai Han, Shihao Zhao, Kwan-Yee K. Wong

Personalized text-to-image generation using diffusion models has recently emerged and garnered significant interest.

Text to Image Generation Text-to-Image Generation

Uni-ControlNet: All-in-One Control to Text-to-Image Diffusion Models

1 code implementation NeurIPS 2023 Shihao Zhao, Dongdong Chen, Yen-Chun Chen, Jianmin Bao, Shaozhe Hao, Lu Yuan, Kwan-Yee K. Wong

Text-to-Image diffusion models have made tremendous progress over the past two years, enabling the generation of highly realistic images based on open-domain text descriptions.

All

CiPR: An Efficient Framework with Cross-instance Positive Relations for Generalized Category Discovery

1 code implementation14 Apr 2023 Shaozhe Hao, Kai Han, Kwan-Yee K. Wong

GCD considers the open-world problem of automatically clustering a partially labelled dataset, in which the unlabelled data may contain instances from both novel categories and labelled classes.

Clustering Contrastive Learning +1

A Unified Framework for Masked and Mask-Free Face Recognition via Feature Rectification

1 code implementation15 Feb 2022 Shaozhe Hao, Chaofeng Chen, Zhenfang Chen, Kwan-Yee K. Wong

We introduce rectification blocks to rectify features extracted by a state-of-the-art recognition model, in both spatial and channel dimensions, to minimize the distance between a masked face and its mask-free counterpart in the rectified feature space.

Face Recognition

Cannot find the paper you are looking for? You can Submit a new open access paper.