Search Results for author: Qingrong Cheng

Found 4 papers, 1 papers with code

Vision-Language Matching for Text-to-Image Synthesis via Generative Adversarial Networks

no code implementations20 Aug 2022 Qingrong Cheng, Keyu Wen, Xiaodong Gu

To address this issue, we propose a novel Vision-Language Matching strategy for text-to-image synthesis, named VLMGAN*, which introduces a dual vision-language matching mechanism to strengthen the image quality and semantic consistency.

Image Generation

Contrastive Cross-Modal Knowledge Sharing Pre-training for Vision-Language Representation Learning and Retrieval

no code implementations2 Jul 2022 Keyu Wen, Zhenshan Tan, Qingrong Cheng, Cheng Chen, Xiaodong Gu

Concretely, the first module is a weight-sharing transformer that builds on the head of the visual and textual encoders, aiming to semantically align text and image.

Contrastive Learning Cross-Modal Retrieval +5

UTC: A Unified Transformer with Inter-Task Contrastive Learning for Visual Dialog

no code implementations CVPR 2022 Cheng Chen, Yudong Zhu, Zhenshan Tan, Qingrong Cheng, Xin Jiang, Qun Liu, Xiaodong Gu

In this paper, we propose a contrastive learning-based framework UTC to unify and facilitate both discriminative and generative tasks in visual dialog with a single model.

Contrastive Learning Representation Learning +1

Learning Dual Semantic Relations with Graph Attention for Image-Text Matching

1 code implementation22 Oct 2020 Keyu Wen, Xiaodong Gu, Qingrong Cheng

Thus, a novel multi-level semantic relations enhancement approach named Dual Semantic Relations Attention Network(DSRAN) is proposed which mainly consists of two modules, separate semantic relations module and the joint semantic relations module.

Graph Attention Image-text matching +1

Cannot find the paper you are looking for? You can Submit a new open access paper.