Search Results for author: Benzhi Wang

Found 5 papers, 2 papers with code

RealisHuman: A Two-Stage Approach for Refining Malformed Human Parts in Generated Images

1 code implementation5 Sep 2024 Benzhi Wang, Jingkai Zhou, Jingqi Bai, Yang Yang, Weihua Chen, Fan Wang, Zhen Lei

First, it generates realistic human parts, such as hands or faces, using the original malformed parts as references, ensuring consistent details with the original image.

Self-similarity Driven Scale-invariant Learning for Weakly Supervised Person Search

no code implementations ICCV 2023 Benzhi Wang, Yang Yang, Jinlin Wu, Guo-Jun Qi, Zhen Lei

On the other hand, the similarity of cross-scale images is often smaller than that of images with the same scale for a person, which will increase the difficulty of matching.

Person Search

Cross-media Scientific Research Achievements Query based on Ranking Learning

no code implementations26 Apr 2022 Benzhi Wang, Meiyu Liang, Ang Li

With the advent of the information age, the scale of data on the Internet is getting larger and larger, and it is full of text, images, videos, and other information.

Decision Making

Cross-Media Scientific Research Achievements Retrieval Based on Deep Language Model

no code implementations29 Mar 2022 Benzhi Wang, Meiyu Liang, Feifei Kou, Mingying Xu

Science and technology big data contain a lot of cross-media information. There are images and texts in the scientific paper. The s ingle modal search method cannot well meet the needs of scientific researchers. This paper proposes a cross-media scientific research achievements retrieval method based on deep language model (CARDL). It achieves a unified cross-media semantic representation by learning the semantic association between different modal data, and is applied to the generation of text semantic vector of scientific research achievements, and then cross-media retrieval is realized through semantic similarity matching between different modal data. Experimental results show that the proposed CARDL method achieves better cross-modal retrieval performance than existing methods.

Cross-Modal Retrieval Language Modelling +3

Cannot find the paper you are looking for? You can Submit a new open access paper.