Search Results for author: Xiao-Yong Wei

Found 13 papers, 7 papers with code

Multi-Level Querying using A Knowledge Pyramid

no code implementations31 Jul 2024 Rubing Chen, Xulu Zhang, Jiaxin Wu, Wenqi Fan, Xiao-Yong Wei, Qing Li

We propose a multi-layer knowledge pyramid approach within the RAG framework to achieve a better balance between precision and recall.

Knowledge Graphs RAG +1

Generative Active Learning for Image Synthesis Personalization

1 code implementation22 Mar 2024 Xulu Zhang, WengYu Zhang, Xiao-Yong Wei, Jinlin Wu, Zhaoxiang Zhang, Zhen Lei, Qing Li

The primary challenge in conducting active learning on generative models lies in the open-ended nature of querying, which differs from the closed form of querying in discriminative models that typically target a single concept.

Active Learning Image Generation

A Picture Is Worth a Graph: A Blueprint Debate Paradigm for Multimodal Reasoning

1 code implementation22 Mar 2024 Changmeng Zheng, Dayong Liang, WengYu Zhang, Xiao-Yong Wei, Tat-Seng Chua, Qing Li

The study addresses two key challenges: the trivialization of opinions resulting from excessive summarization and the diversion of focus caused by distractor concepts introduced from images.

Multimodal Reasoning

Instruct Once, Chat Consistently in Multiple Rounds: An Efficient Tuning Framework for Dialogue

1 code implementation10 Feb 2024 Jian Wang, Chak Tou Leong, Jiashuo Wang, Dongding Lin, Wenjie Li, Xiao-Yong Wei

Tuning language models for dialogue generation has been a prevalent paradigm for building capable dialogue agents.

Dialogue Generation

Compositional Inversion for Stable Diffusion Models

1 code implementation13 Dec 2023 Xulu Zhang, Xiao-Yong Wei, Jinlin Wu, Tianyi Zhang, Zhaoxiang Zhang, Zhen Lei, Qing Li

It stems from the fact that during inversion, the irrelevant semantics in the user images are also encoded, forcing the inverted concepts to occupy locations far from the core distribution in the embedding space.

Multi-agent Attacks for Black-box Social Recommendations

no code implementations13 Nov 2023 Shijie Wang, Wenqi Fan, Xiao-Yong Wei, Xiaowei Mei, Shanru Lin, Qing Li

To perform untargeted attacks on social recommender systems, attackers can construct malicious social relationships for fake users to enhance the attack performance.

Decision Making Multi-agent Reinforcement Learning +1

Empowering Molecule Discovery for Molecule-Caption Translation with Large Language Models: A ChatGPT Perspective

1 code implementation11 Jun 2023 Jiatong Li, Yunqing Liu, Wenqi Fan, Xiao-Yong Wei, Hui Liu, Jiliang Tang, Qing Li

In this work, we propose a novel LLM-based framework (MolReGPT) for molecule-caption translation, where an In-Context Few-Shot Molecule Learning paradigm is introduced to empower molecule discovery with LLMs like ChatGPT to perform their in-context learning capability without domain-specific pre-training and fine-tuning.

In-Context Learning Molecule Captioning +3

Large-scale Multi-Modal Pre-trained Models: A Comprehensive Survey

1 code implementation20 Feb 2023 Xiao Wang, Guangyao Chen, Guangwu Qian, Pengcheng Gao, Xiao-Yong Wei, YaoWei Wang, Yonghong Tian, Wen Gao

We also give visualization and analysis of the model parameters and results on representative downstream tasks.

Survey

Conceptor Learning for Class Activation Mapping

no code implementations21 Jan 2022 Guangwu Qian, Zhen-Qun Yang, Xu-Lu Zhang, YaoWei Wang, Qing Li, Xiao-Yong Wei

Class Activation Mapping (CAM) has been widely adopted to generate saliency maps which provides visual explanations for deep neural networks (DNNs).

Relation

Attention on Attention for Image Captioning

5 code implementations ICCV 2019 Lun Huang, Wenmin Wang, Jie Chen, Xiao-Yong Wei

In this paper, we propose an Attention on Attention (AoA) module, which extends the conventional attention mechanisms to determine the relevance between attention results and queries.

Decoder Image Captioning

ParNet: Position-aware Aggregated Relation Network for Image-Text matching

no code implementations17 Jun 2019 Yaxian Xia, Lun Huang, Xiao-Yong Wei, Wenmin Wang

The first step, we call it intra-modal relation mechanism, in which we computes responses between different objects in an image or different words in a sentence separately; The second step, we call it inter-modal relation mechanism, in which the query plays a role of textual context to refine the relationship among object proposals in an image.

Image-text matching Position +5

Cannot find the paper you are looking for? You can Submit a new open access paper.