Search Results for author: Zhaoqing Wang

Found 16 papers, 6 papers with code

MF-VITON: High-Fidelity Mask-Free Virtual Try-On with Minimal Input

no code implementations11 Mar 2025 Zhenchen Wan, Yanwu Xu, Dongting Hu, Weilun Cheng, Tianxi Chen, Zhaoqing Wang, Feng Liu, Tongliang Liu, Mingming Gong

To address this, we propose a Mask-Free VITON (MF-VITON) framework that achieves realistic VITON using only a single person image and a target garment, eliminating the requirement for auxiliary masks.

Virtual Try-on

OVGaussian: Generalizable 3D Gaussian Segmentation with Open Vocabularies

no code implementations31 Dec 2024 Runnan Chen, Xiangyu Sun, Zhaoqing Wang, Youquan Liu, Jiepeng Wang, Lingdong Kong, Jiankang Deng, Mingming Gong, Liang Pan, Wenping Wang, Tongliang Liu

We first construct a large-scale 3D scene dataset based on 3DGS, dubbed \textbf{SegGaussian}, which provides detailed semantic and instance annotations for both Gaussian points and multi-view images.

3DGS 3D Semantic Segmentation +3

TED-VITON: Transformer-Empowered Diffusion Models for Virtual Try-On

1 code implementation26 Nov 2024 Zhenchen Wan, Yanwu Xu, Zhaoqing Wang, Feng Liu, Tongliang Liu, Mingming Gong

Recent advancements in Virtual Try-On (VTO) have demonstrated exceptional efficacy in generating realistic images and preserving garment details, largely attributed to the robust generative capabilities of text-to-image (T2I) diffusion backbones.

Large Language Model Text Generation +1

Beyond Gaussians: Fast and High-Fidelity 3D Splatting with Linear Kernels

no code implementations19 Nov 2024 Haodong Chen, Runnan Chen, Qiang Qu, Zhaoqing Wang, Tongliang Liu, Xiaoming Chen, Yuk Ying Chung

Recent advancements in 3D Gaussian Splatting (3DGS) have substantially improved novel view synthesis, enabling high-quality reconstruction and real-time rendering.

3DGS Novel View Synthesis

LaVin-DiT: Large Vision Diffusion Transformer

no code implementations18 Nov 2024 Zhaoqing Wang, Xiaobo Xia, Runnan Chen, Dongdong Yu, Changhu Wang, Mingming Gong, Tongliang Liu

Second, for generative modeling, we develop a joint diffusion transformer that progressively produces vision outputs.

In-Context Learning

Training-Free Robust Interactive Video Object Segmentation

no code implementations8 Jun 2024 Xiaoli Wei, Zhaoqing Wang, Yandong Guo, Chunxia Zhang, Tongliang Liu, Mingming Gong

In this paper, we propose a training-free prompt tracking framework for interactive video object segmentation (I-PT), leveraging the powerful generalization of SAM.

Interactive Video Object Segmentation Object +6

Open-Vocabulary Segmentation with Unpaired Mask-Text Supervision

2 code implementations14 Feb 2024 Zhaoqing Wang, Xiaobo Xia, Ziye Chen, Xiao He, Yandong Guo, Mingming Gong, Tongliang Liu

While existing weakly-supervised approaches leverage image-text pairs to reduce the expansive annotation cost, the lack of mask supervision makes it difficult for the model to locate multiple instances and accurately group pixels with similar semantics, significantly hampering versatility and performance.

Language Modelling Segmentation +1

IDEAL: Influence-Driven Selective Annotations Empower In-Context Learners in Large Language Models

1 code implementation16 Oct 2023 Shaokun Zhang, Xiaobo Xia, Zhaoqing Wang, Ling-Hao Chen, Jiale Liu, Qingyun Wu, Tongliang Liu

However, since the prompts need to be sampled from a large volume of annotated examples, finding the right prompt may result in high annotation costs.

In-Context Learning

Mosaic Representation Learning for Self-supervised Visual Pre-training

2 code implementations journal 2023 Zhaoqing Wang, Ziyu Chen, Yaqian Li, Yandong Guo, Jun Yu, Mingming Gong, Tongliang Liu

To address this problem, we propose a mosaic representation learning framework (MosRep), consisting of a new data augmentation strategy that enriches the backgrounds of each small crop and improves the quality of visual representations.

Data Augmentation Representation Learning +1

BEV-SAN: Accurate BEV 3D Object Detection via Slice Attention Networks

no code implementations CVPR 2023 Xiaowei Chi, Jiaming Liu, Ming Lu, Rongyu Zhang, Zhaoqing Wang, Yandong Guo, Shanghang Zhang

In order to find them, we further propose a LiDAR-guided sampling strategy to leverage the statistical distribution of LiDAR to determine the heights of local slices.

3D Object Detection Autonomous Driving +1

CRIS: CLIP-Driven Referring Image Segmentation

1 code implementation CVPR 2022 Zhaoqing Wang, Yu Lu, Qiang Li, Xunqiang Tao, Yandong Guo, Mingming Gong, Tongliang Liu

In addition, we present text-to-pixel contrastive learning to explicitly enforce the text feature similar to the related pixel-level features and dissimilar to the irrelevances.

Contrastive Learning Decoder +4

Exploring Set Similarity for Dense Self-supervised Representation Learning

no code implementations CVPR 2022 Zhaoqing Wang, Qiang Li, Guoxin Zhang, Pengfei Wan, Wen Zheng, Nannan Wang, Mingming Gong, Tongliang Liu

By considering the spatial correspondence, dense self-supervised representation learning has achieved superior performance on various dense prediction tasks.

Instance Segmentation Keypoint Detection +5

Contextual Graph Reasoning Networks

no code implementations1 Jan 2021 Zhaoqing Wang, Jiaming Liu, Yangyuxuan Kang, Mingming Gong, Chuang Zhang, Ming Lu, Ming Wu

Graph Reasoning has shown great potential recently in modeling long-range dependencies, which are crucial for various computer vision tasks.

2D Human Pose Estimation Instance Segmentation +4

Cannot find the paper you are looking for? You can Submit a new open access paper.